var/home/core/zuul-output/0000755000175000017500000000000015114311203014514 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114322707015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005542775615114322701017716 0ustar rootrootDec 04 13:57:21 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 13:57:21 crc restorecon[4693]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:21 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:57:22 crc restorecon[4693]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 13:57:23 crc kubenswrapper[4715]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:57:23 crc kubenswrapper[4715]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 13:57:23 crc kubenswrapper[4715]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:57:23 crc kubenswrapper[4715]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:57:23 crc kubenswrapper[4715]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 13:57:23 crc kubenswrapper[4715]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.033195 4715 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036132 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036151 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036155 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036161 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036166 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036170 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036175 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036181 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036186 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036191 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036195 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036200 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036203 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036207 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036211 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036215 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036218 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036222 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036232 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036236 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036240 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036244 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036247 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036251 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036254 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036258 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036261 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036265 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036268 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036272 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036276 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036279 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036282 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036286 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036289 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036293 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036296 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036300 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036303 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036307 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036311 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036314 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036318 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036321 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036325 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036328 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036332 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036336 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036339 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036343 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036346 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036350 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036353 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036357 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036360 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036364 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036368 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036371 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036375 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036379 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036384 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036389 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036392 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036396 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036399 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036403 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036407 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036412 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036415 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036419 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.036422 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036516 4715 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036531 4715 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036542 4715 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036549 4715 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036556 4715 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036562 4715 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036569 4715 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036575 4715 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036580 4715 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036585 4715 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036591 4715 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036597 4715 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036602 4715 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036607 4715 flags.go:64] FLAG: --cgroup-root="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036612 4715 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036617 4715 flags.go:64] FLAG: --client-ca-file="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036621 4715 flags.go:64] FLAG: --cloud-config="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036625 4715 flags.go:64] FLAG: --cloud-provider="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036630 4715 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036635 4715 flags.go:64] FLAG: --cluster-domain="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036639 4715 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036644 4715 flags.go:64] FLAG: --config-dir="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036648 4715 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036653 4715 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036659 4715 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036664 4715 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036668 4715 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036672 4715 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036677 4715 flags.go:64] FLAG: --contention-profiling="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036681 4715 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036685 4715 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036689 4715 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036693 4715 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036699 4715 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036703 4715 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036707 4715 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036711 4715 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036715 4715 flags.go:64] FLAG: --enable-server="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036719 4715 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036725 4715 flags.go:64] FLAG: --event-burst="100" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036730 4715 flags.go:64] FLAG: --event-qps="50" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036735 4715 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036744 4715 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036752 4715 flags.go:64] FLAG: --eviction-hard="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036759 4715 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036764 4715 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036769 4715 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036774 4715 flags.go:64] FLAG: --eviction-soft="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036779 4715 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036783 4715 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036788 4715 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036794 4715 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036799 4715 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036803 4715 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036807 4715 flags.go:64] FLAG: --feature-gates="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036814 4715 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036818 4715 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036823 4715 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036827 4715 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036832 4715 flags.go:64] FLAG: --healthz-port="10248" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036836 4715 flags.go:64] FLAG: --help="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036840 4715 flags.go:64] FLAG: --hostname-override="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036844 4715 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036848 4715 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036852 4715 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036856 4715 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036860 4715 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036864 4715 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036868 4715 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036873 4715 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036877 4715 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036881 4715 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036885 4715 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036889 4715 flags.go:64] FLAG: --kube-reserved="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036893 4715 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036897 4715 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036901 4715 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036905 4715 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036909 4715 flags.go:64] FLAG: --lock-file="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036914 4715 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036918 4715 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036922 4715 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036929 4715 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036934 4715 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036938 4715 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036942 4715 flags.go:64] FLAG: --logging-format="text" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036947 4715 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036952 4715 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036957 4715 flags.go:64] FLAG: --manifest-url="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036961 4715 flags.go:64] FLAG: --manifest-url-header="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036966 4715 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036971 4715 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036977 4715 flags.go:64] FLAG: --max-pods="110" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036981 4715 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036986 4715 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036990 4715 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036994 4715 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.036998 4715 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037002 4715 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037006 4715 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037015 4715 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037019 4715 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037023 4715 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037027 4715 flags.go:64] FLAG: --pod-cidr="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037049 4715 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037056 4715 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037060 4715 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037064 4715 flags.go:64] FLAG: --pods-per-core="0" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037069 4715 flags.go:64] FLAG: --port="10250" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037073 4715 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037077 4715 flags.go:64] FLAG: --provider-id="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037081 4715 flags.go:64] FLAG: --qos-reserved="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037085 4715 flags.go:64] FLAG: --read-only-port="10255" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037089 4715 flags.go:64] FLAG: --register-node="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037093 4715 flags.go:64] FLAG: --register-schedulable="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037098 4715 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037105 4715 flags.go:64] FLAG: --registry-burst="10" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037109 4715 flags.go:64] FLAG: --registry-qps="5" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037113 4715 flags.go:64] FLAG: --reserved-cpus="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037117 4715 flags.go:64] FLAG: --reserved-memory="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037122 4715 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037126 4715 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037130 4715 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037134 4715 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037138 4715 flags.go:64] FLAG: --runonce="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037143 4715 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037147 4715 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037152 4715 flags.go:64] FLAG: --seccomp-default="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037156 4715 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037160 4715 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037165 4715 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037169 4715 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037173 4715 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037177 4715 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037181 4715 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037185 4715 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037189 4715 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037193 4715 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037198 4715 flags.go:64] FLAG: --system-cgroups="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037202 4715 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037208 4715 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037212 4715 flags.go:64] FLAG: --tls-cert-file="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037216 4715 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037221 4715 flags.go:64] FLAG: --tls-min-version="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037225 4715 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037229 4715 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037233 4715 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037237 4715 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037243 4715 flags.go:64] FLAG: --v="2" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037248 4715 flags.go:64] FLAG: --version="false" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037254 4715 flags.go:64] FLAG: --vmodule="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037259 4715 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.037264 4715 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037368 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037374 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037378 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037382 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037385 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037389 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037392 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037396 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037400 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037403 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037407 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037411 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037415 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037418 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037422 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037425 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037429 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037434 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037438 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037443 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037447 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037451 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037454 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037458 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037462 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037465 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037469 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037473 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037477 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037481 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037486 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037490 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037494 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037497 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037501 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037505 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037508 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037512 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037515 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037518 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037522 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037525 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037529 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037532 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037536 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037540 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037543 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037547 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037550 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037555 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037558 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037562 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037566 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037570 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037573 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037577 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037581 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037585 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037590 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037593 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037597 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037601 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037605 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037610 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037614 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037618 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037623 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037627 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037632 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037636 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.037640 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.038103 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.044743 4715 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.045066 4715 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045131 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045140 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045145 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045149 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045153 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045157 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045161 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045165 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045169 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045172 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045176 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045179 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045183 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045187 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045190 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045194 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045197 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045201 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045205 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045210 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045215 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045219 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045222 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045226 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045231 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045235 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045239 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045244 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045250 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045254 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045258 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045263 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045266 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045272 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045277 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045283 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045288 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045294 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045298 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045303 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045308 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045313 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045318 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045322 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045327 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045331 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045337 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045343 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045347 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045352 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045356 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045360 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045363 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045367 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045371 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045374 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045378 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045382 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045385 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045389 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045392 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045396 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045399 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045403 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045407 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045410 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045414 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045417 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045421 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045425 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045428 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.045434 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045545 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045552 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045556 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045560 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045564 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045567 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045571 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045574 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045578 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045582 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045585 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045589 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045592 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045596 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045599 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045603 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045608 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045612 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045616 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045620 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045623 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045627 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045631 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045634 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045638 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045642 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045647 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045650 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045654 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045657 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045661 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045665 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045668 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045671 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045675 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045679 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045682 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045687 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045690 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045694 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045697 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045701 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045705 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045708 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045712 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045715 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045719 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045722 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045726 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045730 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045733 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045737 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045740 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045744 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045747 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045751 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045754 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045758 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045761 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045765 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045768 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045772 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045776 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045780 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045785 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045789 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045793 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045796 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045800 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045805 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.045809 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.045815 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.045967 4715 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.048915 4715 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.049007 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.049528 4715 server.go:997] "Starting client certificate rotation" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.049547 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.049853 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-23 23:57:56.111259652 +0000 UTC Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.049924 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 466h0m33.061337692s for next certificate rotation Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.055737 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.060343 4715 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.066382 4715 log.go:25] "Validated CRI v1 runtime API" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.082286 4715 log.go:25] "Validated CRI v1 image API" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.084542 4715 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.087582 4715 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-13-53-11-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.087627 4715 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.103465 4715 manager.go:217] Machine: {Timestamp:2025-12-04 13:57:23.101899469 +0000 UTC m=+0.170617704 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:2e286a6a-d778-45b8-8b58-73ec4dd721bd BootID:ab0a07db-261c-49f3-bead-b10436f442e6 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5a:55:88 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5a:55:88 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:91:07:d2 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b8:1d:df Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c4:5a:67 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:29:6c:da Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:9d:2c:06:51:43 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:72:66:76:6e:51:91 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.103733 4715 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.103894 4715 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104180 4715 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104349 4715 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104390 4715 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104620 4715 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104630 4715 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104814 4715 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.104842 4715 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.105278 4715 state_mem.go:36] "Initialized new in-memory state store" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.105363 4715 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.105984 4715 kubelet.go:418] "Attempting to sync node with API server" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.106031 4715 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.106105 4715 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.106120 4715 kubelet.go:324] "Adding apiserver pod source" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.106246 4715 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.109078 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.109099 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.109184 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.109212 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.109691 4715 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.109953 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.110704 4715 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112176 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112219 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112233 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112244 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112261 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112280 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112290 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112305 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112318 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112328 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112342 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.112381 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.113133 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.113198 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.114393 4715 server.go:1280] "Started kubelet" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.115001 4715 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.115005 4715 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.115743 4715 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 13:57:23 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.116697 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e07c42c51dcad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 13:57:23.114359981 +0000 UTC m=+0.183078196,LastTimestamp:2025-12-04 13:57:23.114359981 +0000 UTC m=+0.183078196,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.117700 4715 server.go:460] "Adding debug handlers to kubelet server" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.119448 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.119480 4715 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.119738 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 16:35:43.447568889 +0000 UTC Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.119824 4715 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.119831 4715 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.119852 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.119896 4715 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.120485 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="200ms" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.120607 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.120678 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122205 4715 factory.go:153] Registering CRI-O factory Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122264 4715 factory.go:221] Registration of the crio container factory successfully Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122384 4715 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122400 4715 factory.go:55] Registering systemd factory Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122409 4715 factory.go:221] Registration of the systemd container factory successfully Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122447 4715 factory.go:103] Registering Raw factory Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.122470 4715 manager.go:1196] Started watching for new ooms in manager Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.123218 4715 manager.go:319] Starting recovery of all containers Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.127792 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.127910 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.127982 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128067 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128131 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128188 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128259 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128349 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128432 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128507 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128567 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128621 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128681 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128742 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128797 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128850 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128933 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.128997 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129087 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129145 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129199 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129256 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129310 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129378 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129438 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129507 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129581 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129656 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129789 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129867 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.129955 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130053 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130140 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130198 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130261 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130317 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130374 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130428 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130484 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130566 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130655 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130732 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130863 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.130963 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131074 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131150 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131223 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131298 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131373 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131453 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131529 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131591 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131652 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131711 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131810 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.131915 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132089 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132209 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132319 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132424 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132554 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132661 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132769 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132888 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.132984 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133128 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133237 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133351 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133460 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133574 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133698 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133825 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.133950 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.134084 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.134210 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.134326 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.134444 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.134553 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.135375 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.135460 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136072 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136218 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136240 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136270 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136284 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136300 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136329 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136343 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136364 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136377 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136390 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136413 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136426 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136445 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136461 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136473 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136491 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136505 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136521 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136532 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136543 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136560 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136574 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136589 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136613 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136637 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136658 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136674 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136700 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136739 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136759 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136786 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136812 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136831 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136856 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136870 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136885 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136914 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136929 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136945 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136957 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136969 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136983 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.136996 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137011 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137023 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137052 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137069 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137081 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137095 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137109 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137122 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137136 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137148 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137159 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137172 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137184 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137204 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137215 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137226 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137241 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137253 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137269 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137281 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137295 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137312 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137323 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137337 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137349 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137361 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137376 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137386 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137401 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137414 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137424 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137445 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137459 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137473 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137494 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137514 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137537 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137556 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137570 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137592 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137616 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137635 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137648 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137658 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137673 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137685 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137699 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137711 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137722 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137738 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137748 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137765 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137777 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137789 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137804 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137814 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137828 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137840 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137851 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137867 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137878 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137890 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137909 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137925 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137945 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137964 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137980 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.137997 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.138011 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.138164 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.138213 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.138243 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143257 4715 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143324 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143351 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143372 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143392 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143411 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143429 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143447 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143467 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143484 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143500 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143517 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143533 4715 reconstruct.go:97] "Volume reconstruction finished" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.143544 4715 reconciler.go:26] "Reconciler: start to sync state" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.148246 4715 manager.go:324] Recovery completed Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.161731 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.163657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.163700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.163711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.164792 4715 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.164812 4715 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.164840 4715 state_mem.go:36] "Initialized new in-memory state store" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.176982 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.178622 4715 policy_none.go:49] "None policy: Start" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.179251 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.179344 4715 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.179399 4715 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.179487 4715 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.179918 4715 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.180026 4715 state_mem.go:35] "Initializing new in-memory state store" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.181262 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.181319 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.220502 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.228776 4715 manager.go:334] "Starting Device Plugin manager" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229015 4715 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229055 4715 server.go:79] "Starting device plugin registration server" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229452 4715 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229471 4715 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229694 4715 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229774 4715 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.229787 4715 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.237127 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.280097 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.280178 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.281218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.281263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.281275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.281437 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.281746 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.281882 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282750 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282795 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.282857 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283756 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.283942 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284499 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.284887 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285486 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285519 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.285941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.286112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.286147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.286157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.321743 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="400ms" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.330567 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.331554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.331582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.331592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.331610 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.332080 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345472 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345509 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345604 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345640 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345657 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345736 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345782 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345835 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345879 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345902 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.345950 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.346012 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.449770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450198 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450246 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450272 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450257 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450344 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450375 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.449955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450386 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450417 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450431 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450436 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450451 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450488 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450498 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450519 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450538 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450574 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.450734 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.532534 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.534195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.534238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.534248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.534277 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.534690 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.604530 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.620489 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.631016 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e68dccbe2a9478dc57f5d2b9aa3882e21b4c7f7229fcf0966a7a40ec22a2fea9 WatchSource:0}: Error finding container e68dccbe2a9478dc57f5d2b9aa3882e21b4c7f7229fcf0966a7a40ec22a2fea9: Status 404 returned error can't find the container with id e68dccbe2a9478dc57f5d2b9aa3882e21b4c7f7229fcf0966a7a40ec22a2fea9 Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.636023 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-de0cdad07122ded03d154ed9ea4a64ae2b2ad295feabc6351910a271f7b2eb94 WatchSource:0}: Error finding container de0cdad07122ded03d154ed9ea4a64ae2b2ad295feabc6351910a271f7b2eb94: Status 404 returned error can't find the container with id de0cdad07122ded03d154ed9ea4a64ae2b2ad295feabc6351910a271f7b2eb94 Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.638264 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.645337 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.651359 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.653662 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-878cffbdd4f7bfa65f2f02888e937ecdcc6c53824e17c70d673fad98e8e78be0 WatchSource:0}: Error finding container 878cffbdd4f7bfa65f2f02888e937ecdcc6c53824e17c70d673fad98e8e78be0: Status 404 returned error can't find the container with id 878cffbdd4f7bfa65f2f02888e937ecdcc6c53824e17c70d673fad98e8e78be0 Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.723289 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="800ms" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.849146 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c20a1780f8032d05e1d376d5172e5133ec2436aaf667805504e33b72b0b3c03e WatchSource:0}: Error finding container c20a1780f8032d05e1d376d5172e5133ec2436aaf667805504e33b72b0b3c03e: Status 404 returned error can't find the container with id c20a1780f8032d05e1d376d5172e5133ec2436aaf667805504e33b72b0b3c03e Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.852810 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-88d08ed2f22600abe7cc998f7fa014c28afeeac43ab9ba0341b995cf776306b7 WatchSource:0}: Error finding container 88d08ed2f22600abe7cc998f7fa014c28afeeac43ab9ba0341b995cf776306b7: Status 404 returned error can't find the container with id 88d08ed2f22600abe7cc998f7fa014c28afeeac43ab9ba0341b995cf776306b7 Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.935143 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.936467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.936513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.936530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:23 crc kubenswrapper[4715]: I1204 13:57:23.936570 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.937094 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 04 13:57:23 crc kubenswrapper[4715]: W1204 13:57:23.984459 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:23 crc kubenswrapper[4715]: E1204 13:57:23.984543 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.114019 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.120075 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 15:46:54.497617431 +0000 UTC Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.120122 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 169h49m30.377497736s for next certificate rotation Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.183798 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e68dccbe2a9478dc57f5d2b9aa3882e21b4c7f7229fcf0966a7a40ec22a2fea9"} Dec 04 13:57:24 crc kubenswrapper[4715]: W1204 13:57:24.388666 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:24 crc kubenswrapper[4715]: E1204 13:57:24.388764 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.390164 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88d08ed2f22600abe7cc998f7fa014c28afeeac43ab9ba0341b995cf776306b7"} Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.391381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c20a1780f8032d05e1d376d5172e5133ec2436aaf667805504e33b72b0b3c03e"} Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.392385 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"878cffbdd4f7bfa65f2f02888e937ecdcc6c53824e17c70d673fad98e8e78be0"} Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.393767 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"de0cdad07122ded03d154ed9ea4a64ae2b2ad295feabc6351910a271f7b2eb94"} Dec 04 13:57:24 crc kubenswrapper[4715]: W1204 13:57:24.482642 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:24 crc kubenswrapper[4715]: E1204 13:57:24.482721 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:24 crc kubenswrapper[4715]: E1204 13:57:24.524322 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="1.6s" Dec 04 13:57:24 crc kubenswrapper[4715]: W1204 13:57:24.731868 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:24 crc kubenswrapper[4715]: E1204 13:57:24.731983 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.737829 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.739413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.739456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.739468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:24 crc kubenswrapper[4715]: I1204 13:57:24.739491 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:24 crc kubenswrapper[4715]: E1204 13:57:24.739864 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.115127 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:25 crc kubenswrapper[4715]: E1204 13:57:25.252110 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e07c42c51dcad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 13:57:23.114359981 +0000 UTC m=+0.183078196,LastTimestamp:2025-12-04 13:57:23.114359981 +0000 UTC m=+0.183078196,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.397991 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a" exitCode=0 Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.398082 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.398135 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.399766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.399798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.399809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.400133 4715 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d" exitCode=0 Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.400171 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.400251 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.401126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.401141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.401149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.402980 4715 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02" exitCode=0 Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.403051 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.403062 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.403901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.403927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.403937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.405697 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.405721 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.405731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.405739 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.405744 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.406384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.406405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.406413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.407801 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe" exitCode=0 Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.407822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe"} Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.407868 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.408701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.408724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.408732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.414401 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.415048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.415075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:25 crc kubenswrapper[4715]: I1204 13:57:25.415087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:25 crc kubenswrapper[4715]: W1204 13:57:25.827385 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:25 crc kubenswrapper[4715]: E1204 13:57:25.827475 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.114944 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 04 13:57:26 crc kubenswrapper[4715]: E1204 13:57:26.125703 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="3.2s" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.340812 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.342590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.342628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.342651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.342681 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:26 crc kubenswrapper[4715]: E1204 13:57:26.343148 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.411603 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399"} Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.411681 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55"} Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.413572 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f" exitCode=0 Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.413598 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f"} Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.413680 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.414441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.414477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.414488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.415231 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.415283 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b"} Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.416233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.416255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.416266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.418320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913"} Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.418354 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3"} Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.418358 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.419991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.420009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.420017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:26 crc kubenswrapper[4715]: I1204 13:57:26.911425 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.422419 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92" exitCode=0 Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.422474 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92"} Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.422493 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.423138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.423163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.423172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.425599 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8"} Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.425721 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.426349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.426420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.426430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.428756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c"} Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.428819 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96"} Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.428836 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c"} Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.428772 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.428852 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.428775 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:27 crc kubenswrapper[4715]: I1204 13:57:27.429891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.433285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1"} Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.433352 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.433396 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.433374 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.433436 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.435833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.435855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.435864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.436716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.437357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:28 crc kubenswrapper[4715]: I1204 13:57:28.437398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.439024 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb"} Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.439098 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.439125 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf"} Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.439913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.439943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.439956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.543911 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.545189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.545232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.545250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.545277 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.820875 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.821083 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.821126 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.822519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.822565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.822576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.912108 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 13:57:29 crc kubenswrapper[4715]: I1204 13:57:29.912191 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:57:30 crc kubenswrapper[4715]: I1204 13:57:30.444600 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3"} Dec 04 13:57:30 crc kubenswrapper[4715]: I1204 13:57:30.444643 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670"} Dec 04 13:57:30 crc kubenswrapper[4715]: I1204 13:57:30.444719 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:30 crc kubenswrapper[4715]: I1204 13:57:30.445418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:30 crc kubenswrapper[4715]: I1204 13:57:30.445465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:30 crc kubenswrapper[4715]: I1204 13:57:30.445479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.059630 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.059793 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.059832 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.060942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.061111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.061136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.446716 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.447624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.447653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.447666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.820358 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.820649 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.821954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.821992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:31 crc kubenswrapper[4715]: I1204 13:57:31.822003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.036255 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.036406 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.037624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.037677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.037690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.778013 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.778310 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.779611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.779663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:32 crc kubenswrapper[4715]: I1204 13:57:32.779683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:33 crc kubenswrapper[4715]: E1204 13:57:33.237407 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 13:57:33 crc kubenswrapper[4715]: I1204 13:57:33.907722 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:33 crc kubenswrapper[4715]: I1204 13:57:33.907890 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:33 crc kubenswrapper[4715]: I1204 13:57:33.909587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:33 crc kubenswrapper[4715]: I1204 13:57:33.909655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:33 crc kubenswrapper[4715]: I1204 13:57:33.909673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:34 crc kubenswrapper[4715]: I1204 13:57:34.862097 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:34 crc kubenswrapper[4715]: I1204 13:57:34.862271 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:34 crc kubenswrapper[4715]: I1204 13:57:34.863601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:34 crc kubenswrapper[4715]: I1204 13:57:34.863652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:34 crc kubenswrapper[4715]: I1204 13:57:34.863672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:34 crc kubenswrapper[4715]: I1204 13:57:34.867415 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.063745 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.063930 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.065167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.065235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.065246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.457617 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.458619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.458645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.458657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:35 crc kubenswrapper[4715]: I1204 13:57:35.463273 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:36 crc kubenswrapper[4715]: I1204 13:57:36.460619 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:36 crc kubenswrapper[4715]: I1204 13:57:36.461741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:36 crc kubenswrapper[4715]: I1204 13:57:36.461773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:36 crc kubenswrapper[4715]: I1204 13:57:36.461785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:36 crc kubenswrapper[4715]: W1204 13:57:36.717300 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:57:36 crc kubenswrapper[4715]: I1204 13:57:36.717382 4715 trace.go:236] Trace[1694060938]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:57:26.715) (total time: 10001ms): Dec 04 13:57:36 crc kubenswrapper[4715]: Trace[1694060938]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:57:36.717) Dec 04 13:57:36 crc kubenswrapper[4715]: Trace[1694060938]: [10.00167315s] [10.00167315s] END Dec 04 13:57:36 crc kubenswrapper[4715]: E1204 13:57:36.717402 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:57:37 crc kubenswrapper[4715]: W1204 13:57:37.064679 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:57:37 crc kubenswrapper[4715]: I1204 13:57:37.064759 4715 trace.go:236] Trace[1506952880]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:57:27.063) (total time: 10001ms): Dec 04 13:57:37 crc kubenswrapper[4715]: Trace[1506952880]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:57:37.064) Dec 04 13:57:37 crc kubenswrapper[4715]: Trace[1506952880]: [10.001113067s] [10.001113067s] END Dec 04 13:57:37 crc kubenswrapper[4715]: E1204 13:57:37.064779 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:57:37 crc kubenswrapper[4715]: I1204 13:57:37.114878 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:57:37 crc kubenswrapper[4715]: W1204 13:57:37.236117 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:57:37 crc kubenswrapper[4715]: I1204 13:57:37.236201 4715 trace.go:236] Trace[1433034814]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:57:27.234) (total time: 10001ms): Dec 04 13:57:37 crc kubenswrapper[4715]: Trace[1433034814]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:57:37.236) Dec 04 13:57:37 crc kubenswrapper[4715]: Trace[1433034814]: [10.001339112s] [10.001339112s] END Dec 04 13:57:37 crc kubenswrapper[4715]: E1204 13:57:37.236221 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:57:39 crc kubenswrapper[4715]: E1204 13:57:39.327374 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Dec 04 13:57:39 crc kubenswrapper[4715]: E1204 13:57:39.546926 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 04 13:57:39 crc kubenswrapper[4715]: I1204 13:57:39.821418 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" start-of-body= Dec 04 13:57:39 crc kubenswrapper[4715]: I1204 13:57:39.821480 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" Dec 04 13:57:39 crc kubenswrapper[4715]: I1204 13:57:39.912153 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 13:57:39 crc kubenswrapper[4715]: I1204 13:57:39.912232 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 13:57:40 crc kubenswrapper[4715]: W1204 13:57:40.624891 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:57:40 crc kubenswrapper[4715]: I1204 13:57:40.624984 4715 trace.go:236] Trace[1290153545]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:57:30.623) (total time: 10001ms): Dec 04 13:57:40 crc kubenswrapper[4715]: Trace[1290153545]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:57:40.624) Dec 04 13:57:40 crc kubenswrapper[4715]: Trace[1290153545]: [10.001729134s] [10.001729134s] END Dec 04 13:57:40 crc kubenswrapper[4715]: E1204 13:57:40.625003 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:57:41 crc kubenswrapper[4715]: I1204 13:57:41.734079 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 13:57:41 crc kubenswrapper[4715]: I1204 13:57:41.734145 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 13:57:43 crc kubenswrapper[4715]: I1204 13:57:43.235561 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.118460 4715 apiserver.go:52] "Watching apiserver" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.122168 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.122570 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.123116 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.123254 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.123534 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:44 crc kubenswrapper[4715]: E1204 13:57:44.123598 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:44 crc kubenswrapper[4715]: E1204 13:57:44.123639 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.123832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.123865 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.123984 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:44 crc kubenswrapper[4715]: E1204 13:57:44.124054 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.124738 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.125731 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.125833 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.125888 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.126003 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.126055 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.126137 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.126180 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.126517 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.150295 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.163812 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.176494 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.187969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.196950 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.205081 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.215540 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.221274 4715 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.824838 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.829175 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.833997 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.836566 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.848091 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.857981 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.867176 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.879143 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.889320 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.898306 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.907776 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.918381 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.929599 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.938429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.948578 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:44 crc kubenswrapper[4715]: I1204 13:57:44.960682 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.083263 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.094577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.095823 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.097626 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.103801 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.112826 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.123665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.136311 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.149892 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.163087 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.182160 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.193725 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.205366 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.214856 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.225024 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.235586 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.247721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.256124 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.478071 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.947422 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.948684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.948788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.948850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:45 crc kubenswrapper[4715]: I1204 13:57:45.948917 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:45 crc kubenswrapper[4715]: E1204 13:57:45.951710 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.180211 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.180213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.180373 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.180476 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.180213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.180546 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.718510 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.719171 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.719985 4715 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.757408 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.781114 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.798561 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.814689 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820765 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820801 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820821 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820851 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820866 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820895 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820910 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820928 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820947 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820962 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820977 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.820994 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821046 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821079 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821111 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821158 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821172 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821186 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821199 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821214 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821228 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821277 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821292 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821327 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821343 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821359 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821373 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821388 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821405 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821528 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821546 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821561 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821577 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821593 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821610 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821624 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821642 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821675 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821689 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821704 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821720 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821735 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821750 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821767 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821782 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821797 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821813 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821829 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821843 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821859 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821874 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821893 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821908 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821954 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821973 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821988 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822002 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822019 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822099 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822126 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822144 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822183 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822206 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822226 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822241 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822257 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822277 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822301 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822326 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822342 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822361 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822383 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822405 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822428 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822445 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822468 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822491 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821325 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822515 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822515 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822535 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821341 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822556 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821427 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821426 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822578 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822602 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822624 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822646 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822695 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822716 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822738 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822761 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822784 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822805 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822837 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822858 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822904 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822926 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822946 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822969 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822994 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823018 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823059 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823082 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823104 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823130 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823153 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823175 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823199 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823222 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823247 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823273 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823320 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823344 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823369 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823394 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823419 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823467 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823489 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823515 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823537 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823561 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823613 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823636 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823660 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823685 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823709 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823733 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823785 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823809 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823858 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823879 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823906 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823930 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823953 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824004 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824046 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824070 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824097 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824120 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824144 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821445 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821550 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821668 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821699 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821804 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821859 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821903 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.821946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822015 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822073 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822163 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822182 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822270 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822287 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822341 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822356 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822387 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830021 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822444 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822605 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822750 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822789 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822830 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822891 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830149 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822921 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.822952 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823052 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823090 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823186 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823234 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823322 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823429 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823446 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823461 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823518 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823647 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823708 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823705 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823736 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.823987 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830380 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824084 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830404 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.824165 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:57:47.324146488 +0000 UTC m=+24.392864703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824383 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824504 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824676 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.824865 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830455 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830516 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830595 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830625 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830655 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830651 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830754 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830788 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830824 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830853 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830875 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830921 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830945 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830970 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830995 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831019 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831083 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831107 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831128 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831152 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831177 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831201 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831222 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831243 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831267 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831287 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831305 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831550 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830755 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830810 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825399 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825402 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825484 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825617 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825799 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825816 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825839 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825878 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825942 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826018 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826067 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826069 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826098 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826286 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826301 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826316 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826900 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.826933 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827054 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827085 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827228 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827349 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827363 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827428 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827487 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827698 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827695 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827896 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.827822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.828265 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830968 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.825195 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.830935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831363 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831406 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831854 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.831963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.832316 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.832827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.833172 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.833366 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.833520 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.833847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.834674 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.834685 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.834711 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.834731 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.835110 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.835442 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.835457 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.835621 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.835863 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.835989 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836004 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836214 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836330 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836350 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836404 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836446 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836483 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836745 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.836961 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.837122 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.837914 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838162 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838171 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838336 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838501 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838503 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838515 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838570 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.838725 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.839026 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.839067 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.839224 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.837175 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.839472 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.839978 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840060 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840079 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840109 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840391 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840477 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840561 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841182 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841412 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841500 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841577 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841675 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841777 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841923 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.840974 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.841499 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.843211 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.843280 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.842533 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.843332 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.842002 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.842691 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.842900 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.842964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.842025 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.843900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.843657 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.844308 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.844824 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845097 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845398 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845567 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845579 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845735 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845771 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845804 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.845887 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846988 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847072 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847109 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847129 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847264 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847279 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847292 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847307 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847318 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847328 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847339 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847348 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847357 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847368 4715 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847378 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847438 4715 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847450 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847462 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.846080 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846067 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.847623 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:47.347551902 +0000 UTC m=+24.416270117 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846341 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846517 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846589 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847322 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.847823 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.846622 4715 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.848003 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.848082 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:47.348064896 +0000 UTC m=+24.416783111 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.848590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.848919 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849056 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849078 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849093 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849104 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849142 4715 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849157 4715 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849171 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849181 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849190 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849224 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849230 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849234 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849289 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849299 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849309 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849366 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849380 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849389 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849399 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849411 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849420 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849430 4715 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849443 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849456 4715 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849467 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849477 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.850764 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.851198 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.852416 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.852542 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.854481 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.858545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.859364 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.859382 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.859396 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.849491 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859463 4715 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859477 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859491 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859504 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859516 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859528 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859542 4715 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859555 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859566 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859578 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859592 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859604 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859616 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859629 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859644 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859655 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859669 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859682 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859695 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859707 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859720 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859731 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859743 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859754 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859766 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859747 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859779 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859908 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859922 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859935 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859948 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859960 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859971 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859984 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.859996 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860008 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860020 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860046 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860060 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860072 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860083 4715 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860094 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860105 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860117 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860129 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860141 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860154 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860167 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860180 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860193 4715 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860206 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860219 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860233 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860244 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860256 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860267 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860279 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860290 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860301 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860313 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860325 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860339 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860352 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860364 4715 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860377 4715 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860390 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860401 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860413 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860425 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860436 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860466 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860480 4715 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860491 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860502 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860513 4715 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860524 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860534 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860547 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860559 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860571 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860582 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860595 4715 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860606 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860617 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860629 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860641 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860653 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860665 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860691 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860702 4715 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860713 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860728 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860740 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860753 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860764 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860775 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860788 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860799 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860810 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860822 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860834 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860846 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860859 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860870 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860883 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860895 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860906 4715 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860917 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860929 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860942 4715 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860953 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860970 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860982 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.860993 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861005 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861016 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861027 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861085 4715 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861100 4715 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861112 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861123 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861135 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861147 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861161 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.861172 4715 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.861842 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.861858 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.861868 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.869576 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.869987 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.873112 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.873224 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.873316 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.874351 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.875346 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.875608 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.877478 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.877740 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.877813 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.877935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.878388 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:47.378352523 +0000 UTC m=+24.447070738 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:46 crc kubenswrapper[4715]: E1204 13:57:46.878409 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:47.378402424 +0000 UTC m=+24.447120639 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.878497 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.879240 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.879539 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.879913 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.881619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.881942 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.886053 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.888289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.890104 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.892026 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.897295 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.898216 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.916117 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.922307 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.922567 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.933737 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.945393 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.957539 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961610 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961625 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961638 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961661 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961752 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961770 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961808 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961822 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961836 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961847 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961765 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961858 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961971 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.961987 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962002 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962015 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962050 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962067 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962081 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962093 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962105 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962117 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962128 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962140 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962151 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962162 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962174 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962186 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962197 4715 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.962209 4715 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.966774 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.975847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.985060 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:46 crc kubenswrapper[4715]: I1204 13:57:46.992791 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.001481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.016652 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.026509 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.035406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.043910 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.052484 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.061933 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.070879 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.080755 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.088593 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.137521 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.144078 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.149714 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:57:47 crc kubenswrapper[4715]: W1204 13:57:47.156318 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8796f63b713ec135d862c07aec18ded066b7829e1d8fb33f7e6daf6fb13e4e4d WatchSource:0}: Error finding container 8796f63b713ec135d862c07aec18ded066b7829e1d8fb33f7e6daf6fb13e4e4d: Status 404 returned error can't find the container with id 8796f63b713ec135d862c07aec18ded066b7829e1d8fb33f7e6daf6fb13e4e4d Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.156389 4715 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:57:47 crc kubenswrapper[4715]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Dec 04 13:57:47 crc kubenswrapper[4715]: set -o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Dec 04 13:57:47 crc kubenswrapper[4715]: source /etc/kubernetes/apiserver-url.env Dec 04 13:57:47 crc kubenswrapper[4715]: else Dec 04 13:57:47 crc kubenswrapper[4715]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Dec 04 13:57:47 crc kubenswrapper[4715]: exit 1 Dec 04 13:57:47 crc kubenswrapper[4715]: fi Dec 04 13:57:47 crc kubenswrapper[4715]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Dec 04 13:57:47 crc kubenswrapper[4715]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 04 13:57:47 crc kubenswrapper[4715]: > logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.158652 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.158910 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.160212 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Dec 04 13:57:47 crc kubenswrapper[4715]: W1204 13:57:47.160545 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-0d7c2cd51e5422347c1a8c9955e6cddef4d942dc8978a4ef9a0b6b8525aacd17 WatchSource:0}: Error finding container 0d7c2cd51e5422347c1a8c9955e6cddef4d942dc8978a4ef9a0b6b8525aacd17: Status 404 returned error can't find the container with id 0d7c2cd51e5422347c1a8c9955e6cddef4d942dc8978a4ef9a0b6b8525aacd17 Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.162344 4715 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:57:47 crc kubenswrapper[4715]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Dec 04 13:57:47 crc kubenswrapper[4715]: if [[ -f "/env/_master" ]]; then Dec 04 13:57:47 crc kubenswrapper[4715]: set -o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: source "/env/_master" Dec 04 13:57:47 crc kubenswrapper[4715]: set +o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: fi Dec 04 13:57:47 crc kubenswrapper[4715]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Dec 04 13:57:47 crc kubenswrapper[4715]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Dec 04 13:57:47 crc kubenswrapper[4715]: ho_enable="--enable-hybrid-overlay" Dec 04 13:57:47 crc kubenswrapper[4715]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Dec 04 13:57:47 crc kubenswrapper[4715]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Dec 04 13:57:47 crc kubenswrapper[4715]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Dec 04 13:57:47 crc kubenswrapper[4715]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 04 13:57:47 crc kubenswrapper[4715]: --webhook-cert-dir="/etc/webhook-cert" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --webhook-host=127.0.0.1 \ Dec 04 13:57:47 crc kubenswrapper[4715]: --webhook-port=9743 \ Dec 04 13:57:47 crc kubenswrapper[4715]: ${ho_enable} \ Dec 04 13:57:47 crc kubenswrapper[4715]: --enable-interconnect \ Dec 04 13:57:47 crc kubenswrapper[4715]: --disable-approver \ Dec 04 13:57:47 crc kubenswrapper[4715]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --wait-for-kubernetes-api=200s \ Dec 04 13:57:47 crc kubenswrapper[4715]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --loglevel="${LOGLEVEL}" Dec 04 13:57:47 crc kubenswrapper[4715]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 04 13:57:47 crc kubenswrapper[4715]: > logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.164099 4715 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:57:47 crc kubenswrapper[4715]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Dec 04 13:57:47 crc kubenswrapper[4715]: if [[ -f "/env/_master" ]]; then Dec 04 13:57:47 crc kubenswrapper[4715]: set -o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: source "/env/_master" Dec 04 13:57:47 crc kubenswrapper[4715]: set +o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: fi Dec 04 13:57:47 crc kubenswrapper[4715]: Dec 04 13:57:47 crc kubenswrapper[4715]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Dec 04 13:57:47 crc kubenswrapper[4715]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 04 13:57:47 crc kubenswrapper[4715]: --disable-webhook \ Dec 04 13:57:47 crc kubenswrapper[4715]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --loglevel="${LOGLEVEL}" Dec 04 13:57:47 crc kubenswrapper[4715]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 04 13:57:47 crc kubenswrapper[4715]: > logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.165256 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.184073 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.184822 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.186081 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.186770 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.187895 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.188399 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.189485 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.190442 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.191017 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.191903 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.192385 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.193477 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.194028 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.194563 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.195468 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.195951 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.196867 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.197291 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.197811 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.198791 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.199316 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.200599 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.201151 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.201827 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.202442 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.203009 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.203671 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.204229 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.204893 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.205453 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.206023 4715 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.206183 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.207840 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.208552 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.211955 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.213885 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.215898 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.216832 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.218241 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.219105 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.220274 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.221017 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.222604 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.223942 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.224543 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.225792 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.226655 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.228109 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.228673 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.229336 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.230341 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.230967 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.232145 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.232656 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.365099 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.365174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.365243 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.365255 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:57:48.365228405 +0000 UTC m=+25.433946620 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.365290 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:48.365276506 +0000 UTC m=+25.433994711 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.365317 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.365428 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.365462 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:48.365455511 +0000 UTC m=+25.434173726 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.465731 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.465772 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.465875 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.465888 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.465898 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.465936 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:48.465923768 +0000 UTC m=+25.534641983 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.466172 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.466216 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.466230 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.466300 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:48.466280267 +0000 UTC m=+25.534998542 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.483607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0d7c2cd51e5422347c1a8c9955e6cddef4d942dc8978a4ef9a0b6b8525aacd17"} Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.485543 4715 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:57:47 crc kubenswrapper[4715]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Dec 04 13:57:47 crc kubenswrapper[4715]: if [[ -f "/env/_master" ]]; then Dec 04 13:57:47 crc kubenswrapper[4715]: set -o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: source "/env/_master" Dec 04 13:57:47 crc kubenswrapper[4715]: set +o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: fi Dec 04 13:57:47 crc kubenswrapper[4715]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Dec 04 13:57:47 crc kubenswrapper[4715]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Dec 04 13:57:47 crc kubenswrapper[4715]: ho_enable="--enable-hybrid-overlay" Dec 04 13:57:47 crc kubenswrapper[4715]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Dec 04 13:57:47 crc kubenswrapper[4715]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Dec 04 13:57:47 crc kubenswrapper[4715]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Dec 04 13:57:47 crc kubenswrapper[4715]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 04 13:57:47 crc kubenswrapper[4715]: --webhook-cert-dir="/etc/webhook-cert" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --webhook-host=127.0.0.1 \ Dec 04 13:57:47 crc kubenswrapper[4715]: --webhook-port=9743 \ Dec 04 13:57:47 crc kubenswrapper[4715]: ${ho_enable} \ Dec 04 13:57:47 crc kubenswrapper[4715]: --enable-interconnect \ Dec 04 13:57:47 crc kubenswrapper[4715]: --disable-approver \ Dec 04 13:57:47 crc kubenswrapper[4715]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --wait-for-kubernetes-api=200s \ Dec 04 13:57:47 crc kubenswrapper[4715]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --loglevel="${LOGLEVEL}" Dec 04 13:57:47 crc kubenswrapper[4715]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 04 13:57:47 crc kubenswrapper[4715]: > logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.486465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8796f63b713ec135d862c07aec18ded066b7829e1d8fb33f7e6daf6fb13e4e4d"} Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.487645 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.487898 4715 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:57:47 crc kubenswrapper[4715]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Dec 04 13:57:47 crc kubenswrapper[4715]: if [[ -f "/env/_master" ]]; then Dec 04 13:57:47 crc kubenswrapper[4715]: set -o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: source "/env/_master" Dec 04 13:57:47 crc kubenswrapper[4715]: set +o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: fi Dec 04 13:57:47 crc kubenswrapper[4715]: Dec 04 13:57:47 crc kubenswrapper[4715]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Dec 04 13:57:47 crc kubenswrapper[4715]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 04 13:57:47 crc kubenswrapper[4715]: --disable-webhook \ Dec 04 13:57:47 crc kubenswrapper[4715]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Dec 04 13:57:47 crc kubenswrapper[4715]: --loglevel="${LOGLEVEL}" Dec 04 13:57:47 crc kubenswrapper[4715]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 04 13:57:47 crc kubenswrapper[4715]: > logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.488362 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7665be20ce54388a3fbbe84dade42625a99ff12ac326dbe99bac0ee38c04d898"} Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.488919 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.488974 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.489726 4715 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:57:47 crc kubenswrapper[4715]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Dec 04 13:57:47 crc kubenswrapper[4715]: set -o allexport Dec 04 13:57:47 crc kubenswrapper[4715]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Dec 04 13:57:47 crc kubenswrapper[4715]: source /etc/kubernetes/apiserver-url.env Dec 04 13:57:47 crc kubenswrapper[4715]: else Dec 04 13:57:47 crc kubenswrapper[4715]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Dec 04 13:57:47 crc kubenswrapper[4715]: exit 1 Dec 04 13:57:47 crc kubenswrapper[4715]: fi Dec 04 13:57:47 crc kubenswrapper[4715]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Dec 04 13:57:47 crc kubenswrapper[4715]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 04 13:57:47 crc kubenswrapper[4715]: > logger="UnhandledError" Dec 04 13:57:47 crc kubenswrapper[4715]: E1204 13:57:47.490868 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.495973 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.504352 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.511851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.520117 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.528015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.536062 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.549731 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.558461 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.568432 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.589707 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.599180 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.608003 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.619341 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.626205 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.634050 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.641601 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.649971 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:47 crc kubenswrapper[4715]: I1204 13:57:47.657889 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.179971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.179971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.180052 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.180129 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.180262 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.180382 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.296015 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.373239 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.373332 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.373396 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:57:50.373368176 +0000 UTC m=+27.442086391 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.373452 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.373457 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.373494 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.373564 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:50.373553081 +0000 UTC m=+27.442271416 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.373606 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:50.373586762 +0000 UTC m=+27.442304987 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.474627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:48 crc kubenswrapper[4715]: I1204 13:57:48.474681 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.474799 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.474814 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.474823 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.474871 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:50.47485778 +0000 UTC m=+27.543575995 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.474948 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.475007 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.475019 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:48 crc kubenswrapper[4715]: E1204 13:57:48.475098 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:50.475081666 +0000 UTC m=+27.543799941 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.179672 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.179708 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.179835 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.179884 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.179943 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.180022 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.390284 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.390365 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.390395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.390461 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.390518 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:57:54.390485889 +0000 UTC m=+31.459204104 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.390562 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:54.390552821 +0000 UTC m=+31.459271036 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.390574 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.390656 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:54.390630073 +0000 UTC m=+31.459348318 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.490699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:50 crc kubenswrapper[4715]: I1204 13:57:50.490746 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490853 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490867 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490886 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490895 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490944 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490958 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.490945 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:54.490929545 +0000 UTC m=+31.559647760 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:50 crc kubenswrapper[4715]: E1204 13:57:50.491060 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:57:54.491016478 +0000 UTC m=+31.559734723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.637540 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2r8q9"] Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.638161 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.640672 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.640827 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.640916 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.656839 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-kb882"] Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.657282 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.660423 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.660967 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.662344 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.663868 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.674694 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.687932 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.700749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.704226 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr2tx\" (UniqueName: \"kubernetes.io/projected/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-kube-api-access-tr2tx\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.704294 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wr64\" (UniqueName: \"kubernetes.io/projected/6c81c781-f5fa-4d45-91ff-88a0c057b03f-kube-api-access-6wr64\") pod \"node-resolver-2r8q9\" (UID: \"6c81c781-f5fa-4d45-91ff-88a0c057b03f\") " pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.704324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-host\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.704350 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6c81c781-f5fa-4d45-91ff-88a0c057b03f-hosts-file\") pod \"node-resolver-2r8q9\" (UID: \"6c81c781-f5fa-4d45-91ff-88a0c057b03f\") " pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.704378 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-serviceca\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.713451 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.729390 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.741603 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.753924 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.764770 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.775835 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.793838 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.803423 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.805954 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr2tx\" (UniqueName: \"kubernetes.io/projected/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-kube-api-access-tr2tx\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.806007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-host\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.806066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6c81c781-f5fa-4d45-91ff-88a0c057b03f-hosts-file\") pod \"node-resolver-2r8q9\" (UID: \"6c81c781-f5fa-4d45-91ff-88a0c057b03f\") " pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.806100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wr64\" (UniqueName: \"kubernetes.io/projected/6c81c781-f5fa-4d45-91ff-88a0c057b03f-kube-api-access-6wr64\") pod \"node-resolver-2r8q9\" (UID: \"6c81c781-f5fa-4d45-91ff-88a0c057b03f\") " pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.806133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-serviceca\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.806213 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-host\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.806263 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6c81c781-f5fa-4d45-91ff-88a0c057b03f-hosts-file\") pod \"node-resolver-2r8q9\" (UID: \"6c81c781-f5fa-4d45-91ff-88a0c057b03f\") " pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.807588 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-serviceca\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.815418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.822174 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wr64\" (UniqueName: \"kubernetes.io/projected/6c81c781-f5fa-4d45-91ff-88a0c057b03f-kube-api-access-6wr64\") pod \"node-resolver-2r8q9\" (UID: \"6c81c781-f5fa-4d45-91ff-88a0c057b03f\") " pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.824233 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr2tx\" (UniqueName: \"kubernetes.io/projected/f0c5e3e9-6182-4cbe-969f-ded0f25c43e1-kube-api-access-tr2tx\") pod \"node-ca-kb882\" (UID: \"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\") " pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.826501 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.837785 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.849061 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.866290 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.877083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.886023 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.893820 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.906809 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.918321 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.950532 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2r8q9" Dec 04 13:57:51 crc kubenswrapper[4715]: W1204 13:57:51.962261 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c81c781_f5fa_4d45_91ff_88a0c057b03f.slice/crio-a5099ea36c579c8ef5fa03e0addd47bbe35a9ad357029197b7c4aa97677c2f45 WatchSource:0}: Error finding container a5099ea36c579c8ef5fa03e0addd47bbe35a9ad357029197b7c4aa97677c2f45: Status 404 returned error can't find the container with id a5099ea36c579c8ef5fa03e0addd47bbe35a9ad357029197b7c4aa97677c2f45 Dec 04 13:57:51 crc kubenswrapper[4715]: I1204 13:57:51.967741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kb882" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.179772 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.179785 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.179922 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:52 crc kubenswrapper[4715]: E1204 13:57:52.180056 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:52 crc kubenswrapper[4715]: E1204 13:57:52.180246 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:52 crc kubenswrapper[4715]: E1204 13:57:52.180401 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.437183 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nw55d"] Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.437867 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-tblzf"] Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.438115 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.438332 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8dlbk"] Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.439341 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.440293 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.440459 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.440591 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.441128 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xqzzc"] Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.442807 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.442897 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.460067 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.460525 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.460767 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462311 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462340 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462340 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462341 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462478 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462408 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.462414 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.466899 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.468472 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.469710 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.469937 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.470450 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.471641 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.483560 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.497856 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.501069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2r8q9" event={"ID":"6c81c781-f5fa-4d45-91ff-88a0c057b03f","Type":"ContainerStarted","Data":"95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2"} Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.501127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2r8q9" event={"ID":"6c81c781-f5fa-4d45-91ff-88a0c057b03f","Type":"ContainerStarted","Data":"a5099ea36c579c8ef5fa03e0addd47bbe35a9ad357029197b7c4aa97677c2f45"} Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.502813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kb882" event={"ID":"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1","Type":"ContainerStarted","Data":"975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750"} Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.502844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kb882" event={"ID":"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1","Type":"ContainerStarted","Data":"895627078b0481d99f780317dc3bd3b749c72843ee9d312c0a23355b0709c0f4"} Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.509011 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511063 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-systemd\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511101 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511131 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-node-log\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511153 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-config\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511174 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d5ee7c9-f82d-4848-b21f-e529d843456e-proxy-tls\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511196 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-script-lib\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511220 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6v25\" (UniqueName: \"kubernetes.io/projected/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-kube-api-access-p6v25\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511244 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cnibin\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511319 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-cni-binary-copy\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511373 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-netns\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511415 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d5ee7c9-f82d-4848-b21f-e529d843456e-mcd-auth-proxy-config\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-cni-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511549 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-cnibin\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-os-release\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-multus-certs\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511707 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-system-cni-dir\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-var-lib-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511761 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511781 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-etc-kubernetes\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511803 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511824 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovn-node-metrics-cert\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511844 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-kubelet\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-conf-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511885 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9g4s\" (UniqueName: \"kubernetes.io/projected/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-kube-api-access-k9g4s\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-etc-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-system-cni-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.511970 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6v2w\" (UniqueName: \"kubernetes.io/projected/4d5ee7c9-f82d-4848-b21f-e529d843456e-kube-api-access-j6v2w\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-systemd-units\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-bin\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-netd\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512090 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-daemon-config\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-kubelet\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-cni-bin\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512178 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-netns\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512201 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cni-binary-copy\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512264 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-log-socket\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512298 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-k8s-cni-cncf-io\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512325 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-os-release\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-ovn\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512363 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4d5ee7c9-f82d-4848-b21f-e529d843456e-rootfs\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512403 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-slash\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512424 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-cni-multus\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512449 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-env-overrides\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xxk7\" (UniqueName: \"kubernetes.io/projected/727ec519-5a38-4de0-bc0f-69993f95c5ce-kube-api-access-9xxk7\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512490 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-socket-dir-parent\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.512509 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-hostroot\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.536905 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.554270 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.565389 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.577596 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.589555 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.599675 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.608583 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-kubelet\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613527 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-cni-bin\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613555 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-netns\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613576 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cni-binary-copy\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613599 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-os-release\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613638 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-log-socket\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613660 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-k8s-cni-cncf-io\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613695 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-ovn\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613725 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613745 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4d5ee7c9-f82d-4848-b21f-e529d843456e-rootfs\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613774 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-slash\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-cni-multus\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613817 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-socket-dir-parent\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613837 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-hostroot\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613859 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-env-overrides\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613878 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xxk7\" (UniqueName: \"kubernetes.io/projected/727ec519-5a38-4de0-bc0f-69993f95c5ce-kube-api-access-9xxk7\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-systemd\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613927 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-node-log\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-config\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.613981 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d5ee7c9-f82d-4848-b21f-e529d843456e-proxy-tls\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-script-lib\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614056 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6v25\" (UniqueName: \"kubernetes.io/projected/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-kube-api-access-p6v25\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614080 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cnibin\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614114 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-cni-binary-copy\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-netns\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614192 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d5ee7c9-f82d-4848-b21f-e529d843456e-mcd-auth-proxy-config\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614223 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-system-cni-dir\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-var-lib-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-cni-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614303 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-cnibin\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-os-release\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614323 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-netns\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614332 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-cni-bin\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614348 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-multus-certs\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614849 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-kubelet\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614917 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-multus-certs\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4d5ee7c9-f82d-4848-b21f-e529d843456e-rootfs\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614958 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.614976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-hostroot\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615017 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615026 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-slash\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-etc-kubernetes\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615099 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-conf-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615100 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-cni-binary-copy\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615116 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9g4s\" (UniqueName: \"kubernetes.io/projected/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-kube-api-access-k9g4s\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615108 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cni-binary-copy\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-netns\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615171 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-cnibin\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615176 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-ovn\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615189 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-system-cni-dir\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-node-log\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovn-node-metrics-cert\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-os-release\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-cnibin\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615338 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615343 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-log-socket\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615354 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-etc-kubernetes\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-os-release\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-conf-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615433 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-var-lib-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615482 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-kubelet\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-env-overrides\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615512 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-etc-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-system-cni-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615557 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-cni-multus\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6v2w\" (UniqueName: \"kubernetes.io/projected/4d5ee7c9-f82d-4848-b21f-e529d843456e-kube-api-access-j6v2w\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-run-k8s-cni-cncf-io\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615599 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-cni-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615630 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-socket-dir-parent\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615630 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-config\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615648 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-systemd\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615677 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-bin\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-host-var-lib-kubelet\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615716 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-netd\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615733 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-netd\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-daemon-config\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-system-cni-dir\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615784 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-bin\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-systemd-units\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615843 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-etc-openvswitch\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615864 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-systemd-units\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615934 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d5ee7c9-f82d-4848-b21f-e529d843456e-mcd-auth-proxy-config\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.615951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-script-lib\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.616252 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-multus-daemon-config\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.616311 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.617564 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.619510 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d5ee7c9-f82d-4848-b21f-e529d843456e-proxy-tls\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.619639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovn-node-metrics-cert\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.630843 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.632593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6v25\" (UniqueName: \"kubernetes.io/projected/4ae4f7b7-e122-4bf8-b802-54e882e3d6cc-kube-api-access-p6v25\") pod \"multus-8dlbk\" (UID: \"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\") " pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.633212 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xxk7\" (UniqueName: \"kubernetes.io/projected/727ec519-5a38-4de0-bc0f-69993f95c5ce-kube-api-access-9xxk7\") pod \"ovnkube-node-nw55d\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.633723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6v2w\" (UniqueName: \"kubernetes.io/projected/4d5ee7c9-f82d-4848-b21f-e529d843456e-kube-api-access-j6v2w\") pod \"machine-config-daemon-tblzf\" (UID: \"4d5ee7c9-f82d-4848-b21f-e529d843456e\") " pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.635177 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9g4s\" (UniqueName: \"kubernetes.io/projected/8f8e58aa-e858-46aa-9833-11b2c0bc3c5a-kube-api-access-k9g4s\") pod \"multus-additional-cni-plugins-xqzzc\" (UID: \"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\") " pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.642703 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.655425 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.668278 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.681636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.690811 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.701845 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.712743 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.725164 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.734338 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.747679 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.754540 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.758863 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.762111 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 13:57:52 crc kubenswrapper[4715]: W1204 13:57:52.764642 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod727ec519_5a38_4de0_bc0f_69993f95c5ce.slice/crio-20feddcaa9d707bb1925572f3b50d2844d39195328ac8c35cfc1808eecbddd4b WatchSource:0}: Error finding container 20feddcaa9d707bb1925572f3b50d2844d39195328ac8c35cfc1808eecbddd4b: Status 404 returned error can't find the container with id 20feddcaa9d707bb1925572f3b50d2844d39195328ac8c35cfc1808eecbddd4b Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.770703 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8dlbk" Dec 04 13:57:52 crc kubenswrapper[4715]: W1204 13:57:52.772410 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d5ee7c9_f82d_4848_b21f_e529d843456e.slice/crio-0d03d20dbfdeaede6933065549cbedbe770b525152aeeceffd71a990971cf49c WatchSource:0}: Error finding container 0d03d20dbfdeaede6933065549cbedbe770b525152aeeceffd71a990971cf49c: Status 404 returned error can't find the container with id 0d03d20dbfdeaede6933065549cbedbe770b525152aeeceffd71a990971cf49c Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.777467 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.779850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" Dec 04 13:57:52 crc kubenswrapper[4715]: W1204 13:57:52.785860 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ae4f7b7_e122_4bf8_b802_54e882e3d6cc.slice/crio-a98ac357f24fe5488310b8218556a8f26e132f22418b89b084e247390260a5a2 WatchSource:0}: Error finding container a98ac357f24fe5488310b8218556a8f26e132f22418b89b084e247390260a5a2: Status 404 returned error can't find the container with id a98ac357f24fe5488310b8218556a8f26e132f22418b89b084e247390260a5a2 Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.788705 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: W1204 13:57:52.797564 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f8e58aa_e858_46aa_9833_11b2c0bc3c5a.slice/crio-01597c23489b71340425ccb55fe787c84af5e2f76cc75c670340145d1a5bfa7d WatchSource:0}: Error finding container 01597c23489b71340425ccb55fe787c84af5e2f76cc75c670340145d1a5bfa7d: Status 404 returned error can't find the container with id 01597c23489b71340425ccb55fe787c84af5e2f76cc75c670340145d1a5bfa7d Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.798417 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.810943 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.952298 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.954223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.954260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.954271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.954382 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.964604 4715 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.964881 4715 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.966457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.966515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.966539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.966563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.966578 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:52Z","lastTransitionTime":"2025-12-04T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:52 crc kubenswrapper[4715]: E1204 13:57:52.984873 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.988681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.988736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.988752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.988775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:52 crc kubenswrapper[4715]: I1204 13:57:52.988788 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:52Z","lastTransitionTime":"2025-12-04T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: E1204 13:57:53.001269 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.004879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.004939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.004960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.004990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.005006 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: E1204 13:57:53.015497 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.018838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.018883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.018896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.018916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.018932 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: E1204 13:57:53.030647 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.033496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.033537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.033554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.033577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.033588 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: E1204 13:57:53.043646 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: E1204 13:57:53.043763 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.045708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.045747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.045757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.045772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.045782 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.148072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.148114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.148124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.148141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.148153 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.192688 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.202734 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.211807 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.221245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.230333 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.241944 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.249806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.249831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.249839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.249851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.249861 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.251609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.263856 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.285024 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.301719 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.312662 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.321847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.330104 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.349201 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.351676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.351708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.351718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.351736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.351749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.360834 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.454004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.454370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.454469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.454561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.454625 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.506112 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f8e58aa-e858-46aa-9833-11b2c0bc3c5a" containerID="42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2" exitCode=0 Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.506246 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerDied","Data":"42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.506689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerStarted","Data":"01597c23489b71340425ccb55fe787c84af5e2f76cc75c670340145d1a5bfa7d"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.509057 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerStarted","Data":"bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.509088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerStarted","Data":"a98ac357f24fe5488310b8218556a8f26e132f22418b89b084e247390260a5a2"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.515118 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.515258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.515326 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"0d03d20dbfdeaede6933065549cbedbe770b525152aeeceffd71a990971cf49c"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.517259 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" exitCode=0 Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.517297 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.517322 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"20feddcaa9d707bb1925572f3b50d2844d39195328ac8c35cfc1808eecbddd4b"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.529315 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.538610 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.548204 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.558164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.558561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.558690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.558761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.558844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.558783 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.569133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.586599 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.597939 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.612406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.629394 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.661503 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.663350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.663423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.663436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.663454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.663467 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.676065 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.685597 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.694157 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.704643 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.714851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.724636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.734658 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.744753 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.755477 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.765299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.765340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.765349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.765362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.765371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.767308 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.781564 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.790695 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.798910 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.805913 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.844469 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.868070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.868102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.868115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.868131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.868140 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.891234 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.928371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.967150 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.969858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.969888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.969900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.969915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:53 crc kubenswrapper[4715]: I1204 13:57:53.969925 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:53Z","lastTransitionTime":"2025-12-04T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.008722 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.047901 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.071481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.071508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.071517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.071530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.071539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.173817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.173858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.173869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.173885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.173894 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.180048 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.180113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.180176 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.180234 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.180379 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.180462 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.276098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.276149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.276163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.276181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.276193 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.378466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.378676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.378684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.378697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.378707 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.437187 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.437292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.437360 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:58:02.437331372 +0000 UTC m=+39.506049587 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.437391 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.437459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.437505 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:02.437492496 +0000 UTC m=+39.506210791 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.437561 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.437634 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:02.4376163 +0000 UTC m=+39.506334515 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.480813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.480852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.480861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.480876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.480884 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.523930 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.523976 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.523988 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.523999 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.524010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.524021 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.525558 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f8e58aa-e858-46aa-9833-11b2c0bc3c5a" containerID="be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491" exitCode=0 Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.525624 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerDied","Data":"be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.538400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.538443 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538573 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538590 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538623 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.538566 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538635 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538757 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:02.538740744 +0000 UTC m=+39.607458959 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538594 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538808 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:54 crc kubenswrapper[4715]: E1204 13:57:54.538857 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:02.538843537 +0000 UTC m=+39.607561752 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.548133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.562320 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.572780 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.579160 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.582617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.582651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.582662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.582676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.582686 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.586366 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.599792 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.611212 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.627500 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.638594 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.648952 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.657371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.665930 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.673639 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.682949 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.684542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.684657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.684668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.684683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.684693 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.786986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.787053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.787065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.787083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.787094 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.890283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.890311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.890321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.890335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.890346 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.992424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.992463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.992472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.992489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:54 crc kubenswrapper[4715]: I1204 13:57:54.992498 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:54Z","lastTransitionTime":"2025-12-04T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.094868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.094903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.094967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.094982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.094991 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.197509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.197547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.197558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.197575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.197587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.299991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.300028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.300068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.300087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.300097 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.402843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.402895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.402904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.402918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.402927 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.505506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.505548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.505560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.505582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.505594 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.530558 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f8e58aa-e858-46aa-9833-11b2c0bc3c5a" containerID="28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0" exitCode=0 Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.530609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerDied","Data":"28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.540450 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.551399 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.560699 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.568381 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.577319 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.589338 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.597673 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.608589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.608623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.608633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.608649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.608659 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.609665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.626843 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.636974 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.646494 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.669014 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.679102 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.697849 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.711598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.711628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.711637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.711651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.711661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.712148 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.814196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.814228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.814238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.814252 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.814263 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.916435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.916481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.916684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.916702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:55 crc kubenswrapper[4715]: I1204 13:57:55.916716 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:55Z","lastTransitionTime":"2025-12-04T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.018863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.018900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.018909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.018925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.018937 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.121678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.121721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.121732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.121747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.121758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.180138 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.180180 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.180201 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:56 crc kubenswrapper[4715]: E1204 13:57:56.180290 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:56 crc kubenswrapper[4715]: E1204 13:57:56.180403 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:56 crc kubenswrapper[4715]: E1204 13:57:56.180489 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.224367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.224406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.224417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.224432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.224447 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.326955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.326987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.326996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.327008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.327018 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.429119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.429157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.429168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.429183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.429195 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.531708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.532091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.532104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.532123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.532133 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.537273 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.539573 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f8e58aa-e858-46aa-9833-11b2c0bc3c5a" containerID="26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c" exitCode=0 Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.539620 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerDied","Data":"26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.549414 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.558564 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.568070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.577961 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.588623 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.606489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.617423 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.626632 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.633145 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.633776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.633803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.633812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.633826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.633835 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.639721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.656845 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.667125 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.679110 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.688991 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.698768 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.736972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.737011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.737023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.737054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.737068 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.840874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.840950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.840964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.840983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.840999 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.943958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.944000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.944011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.944028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:56 crc kubenswrapper[4715]: I1204 13:57:56.944053 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:56Z","lastTransitionTime":"2025-12-04T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.047066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.047108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.047122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.047138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.047148 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.149942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.150209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.150281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.150349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.150411 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.253194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.253426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.253506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.253588 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.253696 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.355970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.356004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.356013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.356046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.356058 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.458289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.458322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.458334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.458349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.458360 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.548716 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f8e58aa-e858-46aa-9833-11b2c0bc3c5a" containerID="92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7" exitCode=0 Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.548763 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerDied","Data":"92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.560579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.560619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.560630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.560647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.560660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.574505 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.617356 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.633350 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.642009 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.648457 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.658211 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.663388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.663443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.663456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.663475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.663487 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.673013 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.682451 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.690684 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.704383 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.715500 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.724941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.734286 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.743369 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.756440 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.765099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.765130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.765140 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.765158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.765168 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.867018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.867081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.867092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.867107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.867119 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.969796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.969835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.969847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.969863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:57 crc kubenswrapper[4715]: I1204 13:57:57.969873 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:57Z","lastTransitionTime":"2025-12-04T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.073329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.073759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.073910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.074099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.074238 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.177263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.177309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.177323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.177339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.177349 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.179871 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.179991 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:57:58 crc kubenswrapper[4715]: E1204 13:57:58.180005 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.180261 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:57:58 crc kubenswrapper[4715]: E1204 13:57:58.180437 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:57:58 crc kubenswrapper[4715]: E1204 13:57:58.181145 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.280789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.280831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.280841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.280854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.280864 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.383625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.383678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.383690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.383709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.383722 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.486776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.486831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.486845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.486861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.486872 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.558678 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f8e58aa-e858-46aa-9833-11b2c0bc3c5a" containerID="0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9" exitCode=0 Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.558761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerDied","Data":"0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.573272 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.589545 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.590688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.590735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.590750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.590771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.590783 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.607489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.618357 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.629190 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.641182 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.652842 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.664015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.674589 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.682465 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.692716 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.692794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.692827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.692835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.692850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.692859 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.707103 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.718598 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.734710 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.746076 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.795002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.795054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.795065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.795079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.795088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.897918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.897983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.898001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.898023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:58 crc kubenswrapper[4715]: I1204 13:57:58.898069 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:58Z","lastTransitionTime":"2025-12-04T13:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.001447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.001489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.001503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.001523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.001537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.104363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.104412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.104423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.104443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.104455 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.207015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.207057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.207066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.207080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.207089 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.309135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.309168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.309181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.309197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.309209 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.411210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.411246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.411255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.411270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.411279 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.513804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.513999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.514008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.514024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.514095 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.563380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.563422 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.566504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" event={"ID":"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a","Type":"ContainerStarted","Data":"1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.569868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.570122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.570163 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.579136 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.590726 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.592094 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.604654 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.616503 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.616982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.617103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.617195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.617316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.617399 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.627025 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.637583 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.648538 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.659955 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.675886 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.687846 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.705893 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.716789 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.719223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.719266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.719278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.719296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.719308 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.727753 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.736134 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.743382 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.759998 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.773717 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.783741 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.797080 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.806950 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.822246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.822291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.822300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.822316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.822330 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.833109 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.856451 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.870202 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.884670 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.898625 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.909527 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.919238 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.924334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.924365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.924374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.924386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.924397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:57:59Z","lastTransitionTime":"2025-12-04T13:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.933556 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.945839 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:57:59 crc kubenswrapper[4715]: I1204 13:57:59.961194 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:57:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.027113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.027145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.027155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.027167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.027176 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.129133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.129174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.129188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.129206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.129220 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.180355 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.180425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:00 crc kubenswrapper[4715]: E1204 13:58:00.180666 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:00 crc kubenswrapper[4715]: E1204 13:58:00.180737 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.181110 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:00 crc kubenswrapper[4715]: E1204 13:58:00.181462 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.231734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.231769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.231783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.231800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.231811 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.333396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.333431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.333441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.333454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.333464 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.435970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.436012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.436021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.436070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.436088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.537998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.538053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.538062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.538075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.538083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.574482 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.604942 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.619590 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.633265 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.640336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.640519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.640668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.640812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.640944 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.652085 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.687714 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.709352 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.737125 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.743472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.743527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.743540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.743569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.743582 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.755054 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.769355 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.780647 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.793599 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.811514 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.823856 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.846380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.846413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.846423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.846439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.846450 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.847074 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.859958 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.870626 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:00Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.950138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.950198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.950211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.950229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:00 crc kubenswrapper[4715]: I1204 13:58:00.950240 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:00Z","lastTransitionTime":"2025-12-04T13:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.052111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.052151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.052163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.052179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.052191 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.154346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.154427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.154440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.154461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.154472 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.257395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.257457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.257470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.257491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.257503 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.359299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.359338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.359346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.359361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.359371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.461439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.461705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.461782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.461855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.461917 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.563684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.563718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.563727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.563740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.563751 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.578185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.580182 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/0.log" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.582831 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52" exitCode=1 Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.582883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.583621 4715 scope.go:117] "RemoveContainer" containerID="4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.590181 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.601056 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.612922 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.625756 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.639620 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.654084 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.669271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.669314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.669324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.669340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.669353 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.678473 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.741888 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.772066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.772107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.772117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.772133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.772143 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.778774 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.798198 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.811044 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.832963 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.853248 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.865356 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.874870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.874924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.874937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.874956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.874968 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.882952 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.898454 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.912281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.928313 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.946373 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.971296 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.977095 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.977166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.977182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.977207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.977223 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:01Z","lastTransitionTime":"2025-12-04T13:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:01 crc kubenswrapper[4715]: I1204 13:58:01.989391 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.002829 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:01Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.017554 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.031609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.050164 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"message\\\":\\\"o:140\\\\nI1204 13:58:01.316672 5918 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316710 5918 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316798 5918 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317088 5918 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317171 5918 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:01.317216 5918 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:01.317264 5918 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:58:01.317281 5918 factory.go:656] Stopping watch factory\\\\nI1204 13:58:01.317291 5918 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:01.317299 5918 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:01.317376 5918 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.061536 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.072358 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.079818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.079872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.079884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.079901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.079916 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.083838 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.095170 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.104885 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.179995 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.180078 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.180000 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.180223 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.180492 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.180561 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.181765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.181791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.181801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.181815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.181824 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.284963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.285347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.285362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.285380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.285392 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.387473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.388093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.388123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.388145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.388161 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.490482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.490518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.490529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.490543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.490554 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.529095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.529173 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.529211 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.529284 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.529324 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:18.529312202 +0000 UTC m=+55.598030417 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.529439 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:58:18.529407055 +0000 UTC m=+55.598125280 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.529469 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.529560 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:18.529532928 +0000 UTC m=+55.598251193 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.589714 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/0.log" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.592435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.592483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.592501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.592525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.592545 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.593490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.594082 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.606843 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.619465 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.630383 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.630465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630628 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630652 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630663 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630708 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630724 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630668 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630814 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:18.630791836 +0000 UTC m=+55.699510131 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:02 crc kubenswrapper[4715]: E1204 13:58:02.630898 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:18.630875568 +0000 UTC m=+55.699593863 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.632497 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.654544 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"message\\\":\\\"o:140\\\\nI1204 13:58:01.316672 5918 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316710 5918 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316798 5918 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317088 5918 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317171 5918 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:01.317216 5918 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:01.317264 5918 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:58:01.317281 5918 factory.go:656] Stopping watch factory\\\\nI1204 13:58:01.317291 5918 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:01.317299 5918 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:01.317376 5918 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.670245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.691371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.694693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.694723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.694731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.694747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.694757 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.703588 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.717836 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.732146 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.745644 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.756353 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.767754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.780790 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.796491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.796523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.796531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.796545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.796555 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.798952 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.813209 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:02Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.898044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.898209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.898294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.898378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:02 crc kubenswrapper[4715]: I1204 13:58:02.898486 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:02Z","lastTransitionTime":"2025-12-04T13:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.001719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.001770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.001786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.001811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.001828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.103780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.103828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.103845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.103867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.103882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.197453 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.206510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.206548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.206559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.206576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.206587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.214291 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.227537 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.247814 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.269500 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"message\\\":\\\"o:140\\\\nI1204 13:58:01.316672 5918 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316710 5918 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316798 5918 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317088 5918 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317171 5918 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:01.317216 5918 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:01.317264 5918 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:58:01.317281 5918 factory.go:656] Stopping watch factory\\\\nI1204 13:58:01.317291 5918 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:01.317299 5918 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:01.317376 5918 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.283403 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.302919 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.308389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.308431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.308442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.308459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.308469 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.316608 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.327907 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.339699 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.340117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.340152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.340165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.340184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.340197 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.351887 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.352093 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.355201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.355269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.355282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.355298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.355310 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.363868 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.365445 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.368379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.368422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.368430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.368445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.368454 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.375325 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.379592 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.382507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.382537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.382546 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.382560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.382570 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.385777 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.392478 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.395463 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.395595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.395628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.395638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.395654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.395664 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.405277 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.405386 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.410675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.410710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.410718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.410733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.410742 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.512821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.512881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.512896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.512916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.512929 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.597850 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/1.log" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.598467 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/0.log" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.600692 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec" exitCode=1 Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.600734 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.600793 4715 scope.go:117] "RemoveContainer" containerID="4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.601606 4715 scope.go:117] "RemoveContainer" containerID="5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec" Dec 04 13:58:03 crc kubenswrapper[4715]: E1204 13:58:03.601798 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.613059 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.615126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.615155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.615164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.615177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.615185 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.623424 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.640905 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4639c1ae50c1f40263eed29edb21711b96c23197294a977a72183784997e0a52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"message\\\":\\\"o:140\\\\nI1204 13:58:01.316672 5918 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316710 5918 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.316798 5918 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317088 5918 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:58:01.317171 5918 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:01.317216 5918 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:01.317264 5918 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:58:01.317281 5918 factory.go:656] Stopping watch factory\\\\nI1204 13:58:01.317291 5918 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:01.317299 5918 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:01.317376 5918 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.653394 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.672198 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.687170 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.698754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.712129 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.716831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.716898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.716911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.716954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.716965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.727010 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.741969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.754944 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.770303 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.786712 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.800835 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.816026 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.820293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.820343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.820360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.820405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.820422 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.922965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.923006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.923025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.923059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:03 crc kubenswrapper[4715]: I1204 13:58:03.923071 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:03Z","lastTransitionTime":"2025-12-04T13:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.025175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.025251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.025264 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.025281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.025291 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.128199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.128235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.128243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.128256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.128265 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.180378 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.180416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:04 crc kubenswrapper[4715]: E1204 13:58:04.180569 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:04 crc kubenswrapper[4715]: E1204 13:58:04.180665 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.180421 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:04 crc kubenswrapper[4715]: E1204 13:58:04.180785 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.230290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.230337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.230348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.230364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.230377 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.332843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.332882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.332892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.332905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.332914 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.435197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.435410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.435424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.435440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.435452 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.537767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.537857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.537876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.537906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.537927 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.606900 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/1.log" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.612593 4715 scope.go:117] "RemoveContainer" containerID="5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec" Dec 04 13:58:04 crc kubenswrapper[4715]: E1204 13:58:04.612805 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.629586 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.641117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.641211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.641238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.641277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.641303 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.650981 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.668941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.687110 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.703385 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.718546 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.735190 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.744990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.745080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.745099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.745127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.745151 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.754976 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.773441 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.790061 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz"] Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.790719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.792395 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.793891 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.803645 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.818783 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.834937 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.846235 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.848121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.848188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.848202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.848227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.848244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.858312 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.881730 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.897023 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.913798 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.928012 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.939213 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.950322 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.950725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.950769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.950782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.950797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.950807 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:04Z","lastTransitionTime":"2025-12-04T13:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.953858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmmdq\" (UniqueName: \"kubernetes.io/projected/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-kube-api-access-rmmdq\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.953894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.953913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.953947 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.967240 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:04 crc kubenswrapper[4715]: I1204 13:58:04.985425 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.013513 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.035862 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.053734 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054204 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmmdq\" (UniqueName: \"kubernetes.io/projected/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-kube-api-access-rmmdq\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054488 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054506 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.054533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.055532 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.055722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.061322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.070601 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.076022 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmmdq\" (UniqueName: \"kubernetes.io/projected/6d1a1ca5-196c-44a2-88c7-a4cdccc51c28-kube-api-access-rmmdq\") pod \"ovnkube-control-plane-749d76644c-j9sxz\" (UID: \"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.087899 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.104640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.106770 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: W1204 13:58:05.124429 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d1a1ca5_196c_44a2_88c7_a4cdccc51c28.slice/crio-a83c54e4f700aad95d9c431e69a0ac8640d584a415b20ee56fdf42a73d7911ef WatchSource:0}: Error finding container a83c54e4f700aad95d9c431e69a0ac8640d584a415b20ee56fdf42a73d7911ef: Status 404 returned error can't find the container with id a83c54e4f700aad95d9c431e69a0ac8640d584a415b20ee56fdf42a73d7911ef Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.125962 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.141559 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.158234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.158285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.158302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.158364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.158382 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.162652 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.260987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.261022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.261059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.261074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.261085 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.363461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.363509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.363523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.363544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.363563 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.467495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.467562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.467579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.467606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.467623 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.570801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.570866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.570893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.570916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.570935 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.617749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" event={"ID":"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28","Type":"ContainerStarted","Data":"a83c54e4f700aad95d9c431e69a0ac8640d584a415b20ee56fdf42a73d7911ef"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.673958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.674014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.674058 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.674084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.674099 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.776736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.776787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.776799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.776820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.776834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.879699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.879745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.879758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.879775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.879788 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.930233 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-c9ttc"] Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.930802 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:05 crc kubenswrapper[4715]: E1204 13:58:05.930887 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.956182 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.973618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.982453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.982495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.982506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.982522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.982534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:05Z","lastTransitionTime":"2025-12-04T13:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:05 crc kubenswrapper[4715]: I1204 13:58:05.989919 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.005734 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.022714 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.038356 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.056429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.065584 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf2kh\" (UniqueName: \"kubernetes.io/projected/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-kube-api-access-cf2kh\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.065649 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.071496 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.084757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.084820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.084834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.084860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.084877 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.097315 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.116330 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.129484 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.159789 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.167113 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf2kh\" (UniqueName: \"kubernetes.io/projected/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-kube-api-access-cf2kh\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.167237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.167462 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.167560 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:58:06.667540809 +0000 UTC m=+43.736259024 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.180148 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.180187 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.180226 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.180384 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.180511 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.180685 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.182195 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.185320 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf2kh\" (UniqueName: \"kubernetes.io/projected/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-kube-api-access-cf2kh\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.187754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.187805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.187820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.187843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.187859 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.198933 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.212441 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.230222 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.244412 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.291319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.291362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.291374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.291391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.291403 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.393888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.393937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.393951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.393973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.393987 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.498316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.498363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.498375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.498400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.498415 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.601705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.601759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.601773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.601797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.601812 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.622287 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" event={"ID":"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28","Type":"ContainerStarted","Data":"b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.622334 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" event={"ID":"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28","Type":"ContainerStarted","Data":"3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.624065 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.639684 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.656253 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.672007 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.672137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.672263 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:06 crc kubenswrapper[4715]: E1204 13:58:06.672339 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:58:07.672310147 +0000 UTC m=+44.741028362 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.691129 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.705909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.705962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.705974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.705993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.706008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.708969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.725929 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.741770 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.755109 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.771483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.790310 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.806503 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.808494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.808532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.808542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.808560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.808572 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.824241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.847888 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.864884 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.879418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.904780 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.911685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.911735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.911752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.911772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.911787 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:06Z","lastTransitionTime":"2025-12-04T13:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.923269 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.935972 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.967682 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:06 crc kubenswrapper[4715]: I1204 13:58:06.989481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.006476 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.014878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.014925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.014943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.014966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.014983 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.019959 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.030746 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.050014 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.068179 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.087007 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.101937 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.118632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.118718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.118731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.118757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.118771 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.120291 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.140975 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.157163 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.170208 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.186135 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.200472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.217858 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:07Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.222378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.222742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.222816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.222882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.222955 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.325656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.325730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.325743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.325769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.325784 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.427705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.427749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.427764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.427780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.427792 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.531475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.531527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.531538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.531553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.531565 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.635136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.635190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.635209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.635236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.635257 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.682397 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:07 crc kubenswrapper[4715]: E1204 13:58:07.682682 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:07 crc kubenswrapper[4715]: E1204 13:58:07.682801 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:58:09.68277016 +0000 UTC m=+46.751488415 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.738582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.738650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.738670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.738700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.738729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.842265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.842314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.842333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.842358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.842388 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.945246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.945291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.945305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.945325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:07 crc kubenswrapper[4715]: I1204 13:58:07.945340 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:07Z","lastTransitionTime":"2025-12-04T13:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.047590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.047625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.047636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.047650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.047660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.149461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.149502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.149512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.149528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.149537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.180224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.180245 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.180299 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.180306 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:08 crc kubenswrapper[4715]: E1204 13:58:08.180389 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:08 crc kubenswrapper[4715]: E1204 13:58:08.180521 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:08 crc kubenswrapper[4715]: E1204 13:58:08.180652 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:08 crc kubenswrapper[4715]: E1204 13:58:08.180764 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.252447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.252502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.252518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.252541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.252559 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.355920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.355993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.356016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.356069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.356089 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.458147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.458185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.458194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.458209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.458219 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.561492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.561532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.561544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.561560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.561573 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.664837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.664890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.664898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.664912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.664921 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.767951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.768028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.768112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.768144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.768168 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.871349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.871393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.871402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.871447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.871460 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.974237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.974287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.974304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.974323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:08 crc kubenswrapper[4715]: I1204 13:58:08.974337 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:08Z","lastTransitionTime":"2025-12-04T13:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.076977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.077026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.077051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.077065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.077076 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.180149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.180207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.180225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.180245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.180259 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.285949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.286010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.286026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.286079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.286097 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.388689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.388760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.388774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.388794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.388826 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.490742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.490791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.490803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.490821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.490833 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.593293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.593351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.593361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.593375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.593385 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.695421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.695493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.695506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.695524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.695563 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.701875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:09 crc kubenswrapper[4715]: E1204 13:58:09.701987 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:09 crc kubenswrapper[4715]: E1204 13:58:09.702023 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:58:13.70201019 +0000 UTC m=+50.770728405 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.798749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.798837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.798854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.798880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.798895 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.902017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.902105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.902130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.902160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:09 crc kubenswrapper[4715]: I1204 13:58:09.902183 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:09Z","lastTransitionTime":"2025-12-04T13:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.005916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.005985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.006002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.006029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.006079 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.108690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.108745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.108759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.108782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.108796 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.180267 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.180377 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.180295 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:10 crc kubenswrapper[4715]: E1204 13:58:10.180451 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.180563 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:10 crc kubenswrapper[4715]: E1204 13:58:10.180743 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:10 crc kubenswrapper[4715]: E1204 13:58:10.180895 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:10 crc kubenswrapper[4715]: E1204 13:58:10.181019 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.211316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.211356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.211365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.211378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.211388 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.315499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.315572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.315598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.315628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.315650 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.418280 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.418330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.418345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.418365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.418410 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.522400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.522462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.522485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.522510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.522527 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.625791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.625862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.625877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.625894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.625904 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.728615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.728668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.728717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.728796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.728815 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.831686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.831757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.831770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.831787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.831799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.935181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.935319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.935395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.935427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:10 crc kubenswrapper[4715]: I1204 13:58:10.935490 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:10Z","lastTransitionTime":"2025-12-04T13:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.037664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.037737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.037755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.037781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.037799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.140513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.140617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.140677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.140756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.140784 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.244021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.244087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.244096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.244111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.244120 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.347120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.347165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.347176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.347192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.347204 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.449781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.449829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.449840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.449857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.449866 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.552022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.552143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.552158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.552177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.552190 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.654949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.655019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.655046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.655064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.655095 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.757880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.757946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.757963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.757985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.758001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.861542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.861602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.861619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.861643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.861661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.964467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.964532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.964546 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.964564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:11 crc kubenswrapper[4715]: I1204 13:58:11.964579 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:11Z","lastTransitionTime":"2025-12-04T13:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.067995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.068077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.068100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.068121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.068134 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.171444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.171524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.171546 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.171574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.171592 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.179582 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.179716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.179717 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.179801 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:12 crc kubenswrapper[4715]: E1204 13:58:12.179970 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:12 crc kubenswrapper[4715]: E1204 13:58:12.180310 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:12 crc kubenswrapper[4715]: E1204 13:58:12.180587 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:12 crc kubenswrapper[4715]: E1204 13:58:12.180490 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.274784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.274859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.274876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.274902 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.274920 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.379002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.379073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.379085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.379106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.379117 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.481895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.482008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.482074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.482109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.482132 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.585279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.585392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.585415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.585446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.585465 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.687948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.688002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.688010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.688027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.688058 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.790623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.790678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.790695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.790723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.790740 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.894535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.894573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.894585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.894604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.894618 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.997397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.997446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.997455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.997468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:12 crc kubenswrapper[4715]: I1204 13:58:12.997477 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:12Z","lastTransitionTime":"2025-12-04T13:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.100809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.100873 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.100890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.100917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.100938 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.197182 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.203610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.203679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.203693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.203717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.203731 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.212756 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.228827 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.247890 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.263070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.280570 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.306023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.306092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.306106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.306126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.306140 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.306056 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.325481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.342953 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.358282 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.371064 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.399564 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.408437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.408493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.408508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.408528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.408565 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.423619 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.439785 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.460097 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.478629 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.492144 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.511018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.511085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.511101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.511119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.511132 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.614363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.614414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.614430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.614447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.614460 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.710972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.711065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.711080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.711105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.711120 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.726791 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.733426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.733500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.733521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.733554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.733571 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.746983 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.747205 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.747310 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:58:21.747284972 +0000 UTC m=+58.816003197 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.751912 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.757776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.757818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.757835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.757862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.757879 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.779792 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.784895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.784949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.784961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.784986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.785002 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.802255 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.807921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.807990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.808007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.808056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.808079 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.824394 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:13 crc kubenswrapper[4715]: E1204 13:58:13.824561 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.826730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.826779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.826790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.826813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.826825 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.930480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.930566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.930589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.930618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:13 crc kubenswrapper[4715]: I1204 13:58:13.930639 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:13Z","lastTransitionTime":"2025-12-04T13:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.033794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.033872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.033889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.033914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.033928 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.137697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.137777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.137795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.137826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.137848 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.179765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.179794 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:14 crc kubenswrapper[4715]: E1204 13:58:14.179969 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.179789 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:14 crc kubenswrapper[4715]: E1204 13:58:14.180504 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.180593 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:14 crc kubenswrapper[4715]: E1204 13:58:14.180810 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:14 crc kubenswrapper[4715]: E1204 13:58:14.180963 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.240664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.240740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.240762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.240796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.240820 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.343674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.343757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.343779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.343808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.343826 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.446686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.446728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.446740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.446757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.446768 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.549687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.549747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.549764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.549788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.549805 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.652390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.652423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.652432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.652444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.652454 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.754900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.754938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.754946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.754961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.754969 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.858368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.858439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.858456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.858480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.858496 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.962162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.962222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.962240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.962263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:14 crc kubenswrapper[4715]: I1204 13:58:14.962283 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:14Z","lastTransitionTime":"2025-12-04T13:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.064948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.065641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.065676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.065694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.065705 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.168465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.168502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.168513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.168529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.168539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.273677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.273728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.273751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.273773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.273785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.376508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.376556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.376565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.376578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.376586 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.479270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.479322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.479336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.479355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.479368 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.582723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.582777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.582796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.582819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.582837 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.684949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.684978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.684989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.685004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.685012 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.787458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.787516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.787536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.787560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.787580 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.890658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.890720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.890746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.890775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.890798 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.993920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.993980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.994002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.994036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:15 crc kubenswrapper[4715]: I1204 13:58:15.994099 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:15Z","lastTransitionTime":"2025-12-04T13:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.014702 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.025788 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.032603 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.050833 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.070481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.094375 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.097135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.097200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.097218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.097242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.097260 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.109099 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.120672 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.133524 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.143993 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.154926 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.180122 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.180285 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.180515 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:16 crc kubenswrapper[4715]: E1204 13:58:16.180649 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:16 crc kubenswrapper[4715]: E1204 13:58:16.180725 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:16 crc kubenswrapper[4715]: E1204 13:58:16.180882 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.180906 4715 scope.go:117] "RemoveContainer" containerID="5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.181139 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:16 crc kubenswrapper[4715]: E1204 13:58:16.181374 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.183623 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.199877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.199917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.199928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.199945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.199958 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.203024 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.219203 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.241865 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.257063 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.271597 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.283839 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.298280 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.302629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.302689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.302704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.302726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.302745 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.406876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.406921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.406932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.406949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.406963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.510309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.510349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.510357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.510372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.510385 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.612851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.612888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.612896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.612909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.612918 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.654652 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/1.log" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.656882 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.669076 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.680067 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.693489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.709992 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.714316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.714352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.714361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.714377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.714387 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.724016 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.741177 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.750967 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.759765 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.778378 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.795186 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.805927 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.816404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.816451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.816465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.816486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.816498 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.820712 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.832334 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.843919 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.852772 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.862851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.871838 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.882729 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.918554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.918599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.918609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.918628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:16 crc kubenswrapper[4715]: I1204 13:58:16.918640 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:16Z","lastTransitionTime":"2025-12-04T13:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.020538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.020615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.020638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.020669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.020694 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.123106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.123168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.123182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.123202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.123216 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.225285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.225345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.225371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.225421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.225440 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.327959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.328000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.328008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.328023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.328056 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.430595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.430661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.430678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.430704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.430721 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.535449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.535486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.535497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.535513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.535532 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.637674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.637715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.637725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.637740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.637750 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.660474 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/2.log" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.661078 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/1.log" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.663663 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5" exitCode=1 Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.663698 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.663727 4715 scope.go:117] "RemoveContainer" containerID="5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.664327 4715 scope.go:117] "RemoveContainer" containerID="230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5" Dec 04 13:58:17 crc kubenswrapper[4715]: E1204 13:58:17.664450 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.688276 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.730475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.739419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.739630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.739737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.739841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.739924 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.741726 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.760133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.776788 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.788726 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.805299 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.817100 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.830665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.840931 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.842427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.842456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.842465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.842480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.842490 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.852334 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.865548 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.878450 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.892143 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.903932 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.917091 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.930658 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.943559 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:17Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.945132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.945163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.945176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.945193 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:17 crc kubenswrapper[4715]: I1204 13:58:17.945203 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:17Z","lastTransitionTime":"2025-12-04T13:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.047523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.047568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.047580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.047597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.047608 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.150081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.150117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.150125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.150137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.150147 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.179822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.179891 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.179954 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.179958 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.180073 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.180118 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.180161 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.180208 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.252272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.252305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.252316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.252332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.252354 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.355068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.355122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.355144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.355162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.355174 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.457614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.457649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.457657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.457669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.457677 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.560595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.560628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.560640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.560656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.560667 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.593456 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.593590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.593691 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.593759 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:58:50.593724179 +0000 UTC m=+87.662442454 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.593821 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.593903 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.593912 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:50.593888344 +0000 UTC m=+87.662606599 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.594010 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:50.593986876 +0000 UTC m=+87.662705141 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.662690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.662759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.662772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.662787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.662798 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.668066 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/2.log" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.694252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.694311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694473 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694493 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694507 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694523 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694595 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694612 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694564 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:50.694548376 +0000 UTC m=+87.763266601 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:18 crc kubenswrapper[4715]: E1204 13:58:18.694721 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:58:50.694684839 +0000 UTC m=+87.763403054 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.765077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.765124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.765135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.765152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.765164 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.868515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.868569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.868581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.868597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.868613 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.971536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.971583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.971594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.971613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:18 crc kubenswrapper[4715]: I1204 13:58:18.971626 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:18Z","lastTransitionTime":"2025-12-04T13:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.075451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.075501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.075513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.075534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.075546 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.178616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.178658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.178668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.178684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.178694 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.281544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.281589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.281600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.281619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.281629 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.384509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.384581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.384598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.384624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.384643 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.487262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.487336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.487356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.487387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.487405 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.590328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.590366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.590376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.590391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.590402 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.693139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.693185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.693215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.693237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.693254 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.795930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.795964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.795975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.795992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.796004 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.898622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.898675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.898694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.898715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:19 crc kubenswrapper[4715]: I1204 13:58:19.898731 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:19Z","lastTransitionTime":"2025-12-04T13:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.001827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.001926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.001937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.001954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.001965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.103643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.103711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.103729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.103756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.103774 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.180369 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.180472 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.180393 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.180387 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:20 crc kubenswrapper[4715]: E1204 13:58:20.180545 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:20 crc kubenswrapper[4715]: E1204 13:58:20.180625 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:20 crc kubenswrapper[4715]: E1204 13:58:20.180707 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:20 crc kubenswrapper[4715]: E1204 13:58:20.180840 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.206773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.206815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.206826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.206843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.206854 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.309263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.309346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.309383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.309415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.309438 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.412388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.412439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.412451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.412471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.412487 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.515239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.515284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.515296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.515312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.515325 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.617640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.617697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.617716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.617749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.617766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.719512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.719575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.719596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.719612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.719623 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.821760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.821800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.821809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.821824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.821834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.924340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.924385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.924397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.924415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:20 crc kubenswrapper[4715]: I1204 13:58:20.924427 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:20Z","lastTransitionTime":"2025-12-04T13:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.026812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.026877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.026891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.026909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.026920 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.129333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.129393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.129416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.129439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.129455 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.231202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.231269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.231280 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.231295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.231304 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.333736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.333818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.333842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.333871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.333888 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.436693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.436759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.436768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.436784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.436795 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.539394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.539456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.539476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.539500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.539517 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.641868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.641935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.641946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.641962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.641974 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.745181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.745262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.745286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.745318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.745339 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.820943 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:21 crc kubenswrapper[4715]: E1204 13:58:21.821234 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:21 crc kubenswrapper[4715]: E1204 13:58:21.821352 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:58:37.821314095 +0000 UTC m=+74.890032350 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.848197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.848244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.848256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.848273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.848286 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.951218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.951273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.951291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.951314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:21 crc kubenswrapper[4715]: I1204 13:58:21.951330 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:21Z","lastTransitionTime":"2025-12-04T13:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.053693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.053769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.053786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.053812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.053829 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.157641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.157695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.157712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.157734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.157751 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.181212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:22 crc kubenswrapper[4715]: E1204 13:58:22.181385 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.181484 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:22 crc kubenswrapper[4715]: E1204 13:58:22.181764 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.181967 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:22 crc kubenswrapper[4715]: E1204 13:58:22.182236 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.182433 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:22 crc kubenswrapper[4715]: E1204 13:58:22.182646 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.260264 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.260746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.260910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.260994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.261087 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.363795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.363834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.363846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.363861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.363871 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.465904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.465939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.465950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.465967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.465979 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.568771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.568822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.568834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.568867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.568880 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.671305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.671343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.671353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.671370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.671383 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.773699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.773734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.773775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.773793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.773803 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.875757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.875794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.875806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.875819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.875827 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.978145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.978179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.978187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.978199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:22 crc kubenswrapper[4715]: I1204 13:58:22.978208 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:22Z","lastTransitionTime":"2025-12-04T13:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.080336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.080366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.080374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.080386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.080394 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.182540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.182919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.182961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.182975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.182984 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.197957 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.210429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.229338 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.250673 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.263854 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.276614 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.284975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.285053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.285072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.285101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.285118 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.287808 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.306508 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5455dbc8ce036e4e833487155b50f052ab6866e80063ff13b46c2e822b00e0ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:03Z\\\",\\\"message\\\":\\\"r.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:58:02.380131 6100 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:58:02.380132 6100 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:58:02.380149 6100 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:58:02.380155 6100 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:58:02.380164 6100 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:58:02.380183 6100 factory.go:656] Stopping watch factory\\\\nI1204 13:58:02.380201 6100 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:58:02.380202 6100 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:58:02.380203 6100 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:58:02.380216 6100 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:58:02.380154 6100 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:58:02.380228 6100 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:58:02.380126 6100 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:58:02.380243 6100 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 13:58:02.380273 6100 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nF1204 13:58:02.380334 6100 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.324145 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.338472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.351283 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.363923 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.379458 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.387958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.387996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.388007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.388027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.388054 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.391269 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.403169 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.414813 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.428528 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.445213 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.490814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.490843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.490851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.490864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.490874 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.593744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.593804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.593826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.593854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.593875 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.696500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.696531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.696541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.696562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.696577 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.799247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.799327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.799340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.799387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.799401 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.901843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.901899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.901910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.901927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:23 crc kubenswrapper[4715]: I1204 13:58:23.901938 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:23Z","lastTransitionTime":"2025-12-04T13:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.004442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.004512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.004533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.004557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.004575 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.101978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.102110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.102130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.102188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.102205 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.119431 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.123674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.123711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.123721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.123736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.123749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.137021 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.142086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.142187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.142205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.142261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.142279 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.162407 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.168251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.168296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.168305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.168323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.168333 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.180452 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.180622 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.180668 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.180806 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.180985 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.180803 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.181244 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.181321 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.181438 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.186463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.186541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.186559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.186587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.186606 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.204617 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:24 crc kubenswrapper[4715]: E1204 13:58:24.204860 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.207445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.207510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.207527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.207556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.207573 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.310470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.310516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.310528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.310548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.310559 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.412964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.413009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.413021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.413067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.413085 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.515685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.515759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.515774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.515791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.515801 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.619914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.619975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.619987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.620005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.620021 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.722276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.722313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.722321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.722336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.722346 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.825859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.825910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.825922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.825943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.825957 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.928373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.928417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.928429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.928444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:24 crc kubenswrapper[4715]: I1204 13:58:24.928456 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:24Z","lastTransitionTime":"2025-12-04T13:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.031067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.031126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.031142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.031164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.031180 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.133679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.133745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.133759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.133774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.133785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.236484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.236529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.236552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.236573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.236587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.338683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.338711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.338720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.338733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.338742 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.448143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.448190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.448202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.448219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.448232 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.551323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.551361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.551376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.551395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.551410 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.653752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.653818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.653834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.653856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.653869 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.756342 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.756388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.756404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.756426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.756442 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.859794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.859867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.859888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.859917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.859938 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.962617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.962661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.962672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.962689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:25 crc kubenswrapper[4715]: I1204 13:58:25.962701 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:25Z","lastTransitionTime":"2025-12-04T13:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.065305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.065333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.065341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.065354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.065363 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.167920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.167960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.167971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.167990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.168001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.179832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:26 crc kubenswrapper[4715]: E1204 13:58:26.179962 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.180139 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:26 crc kubenswrapper[4715]: E1204 13:58:26.180201 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.180279 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:26 crc kubenswrapper[4715]: E1204 13:58:26.180511 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.180784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:26 crc kubenswrapper[4715]: E1204 13:58:26.180972 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.271001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.271107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.271129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.271160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.271184 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.373629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.373676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.373689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.373706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.373718 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.476201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.476247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.476258 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.476277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.476290 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.579089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.579129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.579137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.579152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.579162 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.681713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.681755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.681765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.681781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.681791 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.783718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.783761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.783775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.783790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.783803 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.885722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.885762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.885773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.885789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.885800 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.988542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.988585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.988596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.988622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:26 crc kubenswrapper[4715]: I1204 13:58:26.988635 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:26Z","lastTransitionTime":"2025-12-04T13:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.090816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.090867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.090881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.090900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.090912 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.193186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.193259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.193282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.193311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.193333 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.295578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.295638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.295650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.295665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.295676 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.398148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.398193 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.398203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.398218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.398228 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.500864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.500908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.500917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.500932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.500943 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.603518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.603553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.603562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.603577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.603586 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.705524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.705585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.705603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.705628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.705645 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.808629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.808722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.808760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.808786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.808801 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.911501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.911531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.911539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.911554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:27 crc kubenswrapper[4715]: I1204 13:58:27.911563 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:27Z","lastTransitionTime":"2025-12-04T13:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.013801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.013848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.013864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.013878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.013886 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.116234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.116267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.116278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.116293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.116305 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.180445 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.180483 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.180477 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:28 crc kubenswrapper[4715]: E1204 13:58:28.180613 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:28 crc kubenswrapper[4715]: E1204 13:58:28.180818 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.180545 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:28 crc kubenswrapper[4715]: E1204 13:58:28.181117 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:28 crc kubenswrapper[4715]: E1204 13:58:28.181232 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.218848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.218912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.218931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.218958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.218976 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.321083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.321132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.321141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.321160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.321172 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.423666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.423704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.423721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.423740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.423750 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.526181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.526223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.526234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.526251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.526293 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.628300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.628337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.628345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.628361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.628371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.730077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.730164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.730181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.730202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.730218 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.831831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.831891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.831903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.831943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.831957 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.840273 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.840961 4715 scope.go:117] "RemoveContainer" containerID="230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5" Dec 04 13:58:28 crc kubenswrapper[4715]: E1204 13:58:28.841214 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.859399 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.871644 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.881396 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.891671 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.910506 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.925817 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.934673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.934720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.934732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.934751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.934763 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:28Z","lastTransitionTime":"2025-12-04T13:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.938655 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.958625 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.970549 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.982908 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:28 crc kubenswrapper[4715]: I1204 13:58:28.993706 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:28Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.009961 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.020007 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.030858 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.037057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.037105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.037117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.037134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.037148 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.045897 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.057241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.070536 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.082283 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:29Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.138923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.138959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.138985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.138998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.139007 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.241785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.241855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.241868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.241886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.241900 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.343868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.343911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.343920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.343936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.343946 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.447561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.447616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.447629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.447654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.447668 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.551453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.551513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.551527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.551553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.551570 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.654502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.654545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.654558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.654576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.654586 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.758184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.758248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.758265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.758293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.758311 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.861206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.861271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.861285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.861308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.861326 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.963686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.963724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.963737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.963755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:29 crc kubenswrapper[4715]: I1204 13:58:29.963766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:29Z","lastTransitionTime":"2025-12-04T13:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.067125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.067181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.067194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.067215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.067229 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.170305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.170375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.170401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.170437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.170460 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.180523 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.180539 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.180586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.180611 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:30 crc kubenswrapper[4715]: E1204 13:58:30.180678 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:30 crc kubenswrapper[4715]: E1204 13:58:30.180823 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:30 crc kubenswrapper[4715]: E1204 13:58:30.180913 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:30 crc kubenswrapper[4715]: E1204 13:58:30.180971 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.273243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.273281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.273290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.273306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.273319 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.376016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.376109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.376127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.376153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.376172 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.477943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.477972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.477981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.477994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.478002 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.580093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.580169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.580185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.580219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.580244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.683432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.683499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.683515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.683541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.683559 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.785832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.785885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.785905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.785924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.785935 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.889368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.889416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.889426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.889449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.889459 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.992129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.992172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.992181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.992199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:30 crc kubenswrapper[4715]: I1204 13:58:30.992210 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:30Z","lastTransitionTime":"2025-12-04T13:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.094681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.094748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.094762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.094778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.094786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.197557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.197603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.197614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.197634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.197646 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.300609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.300664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.300678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.300698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.300711 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.402520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.402561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.402572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.402588 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.402601 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.505060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.505130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.505148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.505178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.505197 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.607935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.607968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.607977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.607989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.607998 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.710398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.710436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.710445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.710458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.710467 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.813005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.813070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.813143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.813164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.813174 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.915337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.915405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.915419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.915434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:31 crc kubenswrapper[4715]: I1204 13:58:31.915444 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:31Z","lastTransitionTime":"2025-12-04T13:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.018351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.018397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.018410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.018435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.018448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.120506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.120560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.120576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.120599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.120617 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.180369 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.180401 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.180468 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:32 crc kubenswrapper[4715]: E1204 13:58:32.180586 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.180645 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:32 crc kubenswrapper[4715]: E1204 13:58:32.180769 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:32 crc kubenswrapper[4715]: E1204 13:58:32.180846 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:32 crc kubenswrapper[4715]: E1204 13:58:32.180940 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.223122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.223157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.223167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.223181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.223192 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.326358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.326408 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.326425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.326448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.326464 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.429404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.429471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.429488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.429514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.429533 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.531570 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.531607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.531616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.531630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.531638 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.634334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.634371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.634381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.634395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.634404 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.736192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.736234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.736245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.736261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.736272 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.838655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.838683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.838691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.838704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.838714 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.940982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.941022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.941047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.941061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:32 crc kubenswrapper[4715]: I1204 13:58:32.941071 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:32Z","lastTransitionTime":"2025-12-04T13:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.043669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.043711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.043719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.043732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.043742 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.146052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.146097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.146106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.146121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.146132 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.193687 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.203358 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.222610 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.233702 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.249323 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.250245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.250284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.250313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.250328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.250352 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.260217 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.271977 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.284110 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.293469 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.302354 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.320830 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.335345 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.346472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.353174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.353206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.353218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.353236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.353248 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.364520 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.375505 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.386199 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.395296 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.407262 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:33Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.455299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.455335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.455345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.455378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.455388 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.557420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.557483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.557495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.557510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.557521 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.660015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.660150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.660175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.660206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.660229 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.763194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.763240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.763253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.763271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.763284 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.866197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.866247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.866259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.866276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.866287 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.967908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.967948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.967961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.967976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:33 crc kubenswrapper[4715]: I1204 13:58:33.967986 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:33Z","lastTransitionTime":"2025-12-04T13:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.069569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.069612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.069624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.069642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.069654 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.172232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.172309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.172320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.172362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.172375 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.180584 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.180624 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.180684 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.180625 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.180745 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.180849 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.180948 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.181077 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.275111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.275171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.275188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.275215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.275234 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.277080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.277123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.277134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.277148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.277159 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.295624 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:34Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.300203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.300332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.300352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.300394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.300416 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.317828 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:34Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.320872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.320915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.320926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.320946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.320959 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.333340 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:34Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.336421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.336471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.336482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.336500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.336511 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.348699 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:34Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.352451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.352487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.352498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.352521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.352532 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.365731 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:34Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:34 crc kubenswrapper[4715]: E1204 13:58:34.365855 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.377114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.377141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.377151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.377170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.377183 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.479742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.479783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.479794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.479810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.479823 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.582572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.582611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.582619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.582634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.582643 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.685062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.685110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.685122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.685139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.685151 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.787589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.787618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.787628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.787643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.787653 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.889464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.889526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.889537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.889549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.889560 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.992110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.992182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.992206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.992235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:34 crc kubenswrapper[4715]: I1204 13:58:34.992256 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:34Z","lastTransitionTime":"2025-12-04T13:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.095542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.095631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.095706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.095743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.095768 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.198412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.198492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.198515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.198544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.198566 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.301586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.301647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.301658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.301677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.301689 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.404859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.404910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.404922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.404940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.404953 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.507449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.507522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.507538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.507555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.507566 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.609940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.609972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.609980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.609994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.610003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.712296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.712337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.712349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.712364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.712375 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.814991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.815098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.815118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.815143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.815160 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.917708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.917756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.917767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.917784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:35 crc kubenswrapper[4715]: I1204 13:58:35.917797 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:35Z","lastTransitionTime":"2025-12-04T13:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.020136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.020187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.020205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.020235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.020254 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.122516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.122594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.122615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.122639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.122660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.180409 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.180553 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:36 crc kubenswrapper[4715]: E1204 13:58:36.180686 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.180728 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.180819 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:36 crc kubenswrapper[4715]: E1204 13:58:36.180965 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:36 crc kubenswrapper[4715]: E1204 13:58:36.181207 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:36 crc kubenswrapper[4715]: E1204 13:58:36.181310 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.225480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.225521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.225530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.225545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.225556 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.328737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.328786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.328797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.328817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.328829 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.431912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.432030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.432097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.432131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.432155 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.534412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.534460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.534474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.534494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.534508 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.636691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.636758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.636772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.636803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.636819 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.746638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.746688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.746699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.746717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.746728 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.849340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.849387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.849397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.849411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.849420 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.952785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.952836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.952848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.952870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:36 crc kubenswrapper[4715]: I1204 13:58:36.952882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:36Z","lastTransitionTime":"2025-12-04T13:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.056129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.056202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.056218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.056242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.056257 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.158857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.158916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.158930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.158951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.158965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.262289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.262370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.262383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.262426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.262443 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.365618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.365685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.365701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.365725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.365738 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.468528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.468587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.468599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.468616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.468725 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.571892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.571972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.571998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.572077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.572105 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.675826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.675903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.675920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.675952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.675972 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.779518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.779592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.779610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.779638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.779664 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.882735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.882803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.882814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.882827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.882837 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.886159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:37 crc kubenswrapper[4715]: E1204 13:58:37.886302 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:37 crc kubenswrapper[4715]: E1204 13:58:37.886360 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 13:59:09.886345288 +0000 UTC m=+106.955063503 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.986670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.987003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.987025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.987099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:37 crc kubenswrapper[4715]: I1204 13:58:37.987116 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:37Z","lastTransitionTime":"2025-12-04T13:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.090690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.090756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.090795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.090831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.090857 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.179830 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.179876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.179878 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.179854 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:38 crc kubenswrapper[4715]: E1204 13:58:38.179967 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:38 crc kubenswrapper[4715]: E1204 13:58:38.180075 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:38 crc kubenswrapper[4715]: E1204 13:58:38.180211 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:38 crc kubenswrapper[4715]: E1204 13:58:38.180380 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.192401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.192440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.192451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.192467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.192479 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.295202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.295246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.295254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.295268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.295279 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.398519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.398576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.398591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.398610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.398624 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.500894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.500991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.501009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.501082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.501121 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.603856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.603897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.603908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.603923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.603936 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.706760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.707301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.707324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.707354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.707373 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.732215 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/0.log" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.732265 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" containerID="bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e" exitCode=1 Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.732312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerDied","Data":"bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.732760 4715 scope.go:117] "RemoveContainer" containerID="bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.746852 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.760698 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.774969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.788446 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.803769 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.809352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.809390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.809399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.809414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.809426 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.819420 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.830572 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.843647 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.866498 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.882640 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.893076 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.912384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.912423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.912432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.912446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.912456 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:38Z","lastTransitionTime":"2025-12-04T13:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.913314 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.925447 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.936784 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.946636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.957548 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.968176 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:38 crc kubenswrapper[4715]: I1204 13:58:38.979306 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.015200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.015246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.015257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.015272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.015284 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.117701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.117737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.117749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.117766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.117777 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.220448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.220500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.220512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.220528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.220541 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.322793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.322834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.322845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.322858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.322868 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.424791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.424824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.424834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.424851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.424862 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.527267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.527313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.527324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.527345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.527359 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.630472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.630509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.630520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.630537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.630548 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.732687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.732730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.732755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.732776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.732791 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.736451 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/0.log" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.736505 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerStarted","Data":"64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.752132 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.766721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.789817 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.804646 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.815297 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.833439 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.835026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.835079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.835092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.835109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.835122 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.845408 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.856640 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.866404 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.878480 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.890652 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.902960 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.914138 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.925591 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.935179 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.936912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.936954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.936966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.936983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.936992 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:39Z","lastTransitionTime":"2025-12-04T13:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.947484 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.958600 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:39 crc kubenswrapper[4715]: I1204 13:58:39.971170 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.039713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.039760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.039773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.039790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.039802 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.141713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.141773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.141786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.141808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.141825 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.180533 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.180585 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.180585 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:40 crc kubenswrapper[4715]: E1204 13:58:40.180688 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.181329 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:40 crc kubenswrapper[4715]: E1204 13:58:40.181481 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.181934 4715 scope.go:117] "RemoveContainer" containerID="230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5" Dec 04 13:58:40 crc kubenswrapper[4715]: E1204 13:58:40.182113 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:40 crc kubenswrapper[4715]: E1204 13:58:40.182510 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.244896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.244968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.244981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.244999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.245012 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.349149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.349191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.349201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.349217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.349227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.453466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.453502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.453512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.453525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.453534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.556111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.556152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.556164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.556180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.556191 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.658573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.658611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.658622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.658636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.658647 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.741590 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/2.log" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.744427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.744743 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.759152 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.760987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.761104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.761132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.761165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.761190 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.773598 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.785665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.800984 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.813150 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.830052 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.842241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.864478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.864514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.864527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.864542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.864553 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.868019 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.882135 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.900641 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.913569 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.926183 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.945580 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.961605 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.966494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.966528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.966537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.966551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.966561 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:40Z","lastTransitionTime":"2025-12-04T13:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.975847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:40 crc kubenswrapper[4715]: I1204 13:58:40.990567 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.002200 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.014679 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.068446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.068479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.068487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.068503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.068513 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.170913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.171004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.171026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.171081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.171099 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.273617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.273650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.273658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.273671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.273679 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.376148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.376212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.376229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.376249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.376264 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.481867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.481962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.481983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.482051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.482078 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.584993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.585098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.585122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.585149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.585169 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.687864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.687924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.687940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.687963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.687982 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.749850 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/3.log" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.750813 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/2.log" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.754647 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" exitCode=1 Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.754707 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.754765 4715 scope.go:117] "RemoveContainer" containerID="230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.755658 4715 scope.go:117] "RemoveContainer" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" Dec 04 13:58:41 crc kubenswrapper[4715]: E1204 13:58:41.755874 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.768781 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.790141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.790182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.790194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.790209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.790221 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.790158 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230d91672357dd1ad040196ab7f2f013a1597fc18e8f02d7ae421f5a4d98beb5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:17Z\\\",\\\"message\\\":\\\"p: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999966 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1204 13:58:16.999986 6309 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1204 13:58:16.999972 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:17.000013 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1204 13:58:17.000025 6309 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1204 13:58:17.000075 6309 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1204 13:58:16.999565 6309 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1204 13:58:17.000100 6309 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1204 13:58:16.999601 6309 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.804664 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.815114 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.834367 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.846211 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.856197 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.867389 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.877886 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.888129 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.891784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.891822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.891831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.891846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.891855 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.899202 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.909078 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.919508 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.929936 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.940019 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.950554 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.971810 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.994398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.994442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.994454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.994472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.994484 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:41Z","lastTransitionTime":"2025-12-04T13:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:41 crc kubenswrapper[4715]: I1204 13:58:41.998927 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.096023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.096104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.096120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.096141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.096157 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.180216 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.180261 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.180234 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.180234 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:42 crc kubenswrapper[4715]: E1204 13:58:42.180357 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:42 crc kubenswrapper[4715]: E1204 13:58:42.180458 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:42 crc kubenswrapper[4715]: E1204 13:58:42.180538 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:42 crc kubenswrapper[4715]: E1204 13:58:42.180608 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.197925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.197969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.197978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.197994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.198005 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.300126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.300175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.300186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.300204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.300216 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.402894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.402942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.402959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.402984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.403000 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.506366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.506425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.506444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.506467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.506485 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.609003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.609093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.609112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.609137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.609153 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.711681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.711723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.711735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.711754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.711768 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.759365 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/3.log" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.762491 4715 scope.go:117] "RemoveContainer" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" Dec 04 13:58:42 crc kubenswrapper[4715]: E1204 13:58:42.762616 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.777316 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.793607 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.805689 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.814689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.814731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.814742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.814763 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.814774 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.817715 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.828754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.839309 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.849481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.858266 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.867204 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.885459 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.899836 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.910475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.917302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.917349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.917361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.917379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.917390 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:42Z","lastTransitionTime":"2025-12-04T13:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.929499 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.944207 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.957023 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.967198 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.979528 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:42 crc kubenswrapper[4715]: I1204 13:58:42.992479 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.020184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.020227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.020235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.020247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.020258 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.122747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.122777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.122785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.122798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.122806 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.191228 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.192482 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.205192 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.217569 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.225183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.225219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.225227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.225240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.225250 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.228941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.237606 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.249046 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.263361 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.275604 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.287151 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.298320 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.306707 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.316022 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.327368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.327413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.327424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.327443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.327456 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.334428 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.347711 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.357714 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.376418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.389078 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.402209 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.430201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.430248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.430260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.430275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.430287 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.532142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.532174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.532181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.532202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.532210 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.634966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.635008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.635021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.635059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.635072 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.736944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.737000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.737013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.737045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.737061 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.839446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.839519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.839556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.839587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.839609 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.941919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.941957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.941967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.941981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:43 crc kubenswrapper[4715]: I1204 13:58:43.941991 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:43Z","lastTransitionTime":"2025-12-04T13:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.044769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.044816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.044826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.044842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.044852 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.146880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.146926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.146940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.146963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.146975 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.179815 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.179847 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.179982 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.179865 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.179847 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.180107 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.180277 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.180371 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.249256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.249296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.249307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.249324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.249336 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.352250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.352289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.352301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.352317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.352329 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.454578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.454620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.454631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.454647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.454658 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.519847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.519907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.519919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.519937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.519949 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.534507 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.539274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.539308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.539318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.539337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.539347 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.553074 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.559521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.559613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.559643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.559686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.559712 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.577187 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.581997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.582083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.582100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.582123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.582144 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.596338 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.600294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.600334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.600345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.600363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.600374 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.611544 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:44 crc kubenswrapper[4715]: E1204 13:58:44.611659 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.613259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.613329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.613341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.613364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.613377 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.715327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.715374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.715385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.715403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.715415 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.818059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.818101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.818112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.818128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.818138 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.920957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.921017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.921059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.921084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:44 crc kubenswrapper[4715]: I1204 13:58:44.921103 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:44Z","lastTransitionTime":"2025-12-04T13:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.023738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.023775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.023784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.023797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.023806 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.126103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.126175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.126194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.126221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.126238 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.228545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.228592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.228606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.228626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.228637 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.330998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.331052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.331064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.331080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.331090 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.432970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.433196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.433253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.433273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.433287 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.535067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.535118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.535130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.535150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.535162 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.637995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.638075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.638089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.638109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.638121 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.740859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.740937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.740960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.740990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.741012 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.843326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.843368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.843381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.843397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.843409 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.946075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.946114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.946125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.946143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:45 crc kubenswrapper[4715]: I1204 13:58:45.946153 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:45Z","lastTransitionTime":"2025-12-04T13:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.048841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.048879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.048888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.048903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.048915 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.151489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.151532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.151543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.151555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.151563 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.180558 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.180562 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.180619 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.180767 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:46 crc kubenswrapper[4715]: E1204 13:58:46.180824 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:46 crc kubenswrapper[4715]: E1204 13:58:46.180759 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:46 crc kubenswrapper[4715]: E1204 13:58:46.180889 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:46 crc kubenswrapper[4715]: E1204 13:58:46.180932 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.254346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.254369 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.254377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.254389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.254399 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.357789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.357848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.357862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.357884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.357896 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.460815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.460884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.460901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.461420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.461495 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.563811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.563844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.563853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.563866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.563874 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.666717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.666765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.666776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.666798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.666811 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.769509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.769583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.769595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.769613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.769625 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.871712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.871770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.871783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.871804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.871818 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.974057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.974099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.974109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.974131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:46 crc kubenswrapper[4715]: I1204 13:58:46.974143 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:46Z","lastTransitionTime":"2025-12-04T13:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.076556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.076628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.076639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.076653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.076664 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.179444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.179490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.179501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.179517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.179527 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.283883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.283936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.283949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.283972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.283989 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.387962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.388007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.388017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.388056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.388068 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.490241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.490288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.490301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.490319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.490331 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.592711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.592769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.592790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.592812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.592829 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.695482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.695519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.695528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.695544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.695553 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.798053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.798155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.798175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.798202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.798226 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.900875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.900948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.900959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.900977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:47 crc kubenswrapper[4715]: I1204 13:58:47.900989 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:47Z","lastTransitionTime":"2025-12-04T13:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.004028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.004103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.004116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.004133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.004143 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.106561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.106600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.106612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.106629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.106641 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.180513 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:48 crc kubenswrapper[4715]: E1204 13:58:48.180647 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.180526 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.180805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.180845 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:48 crc kubenswrapper[4715]: E1204 13:58:48.181218 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:48 crc kubenswrapper[4715]: E1204 13:58:48.181389 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:48 crc kubenswrapper[4715]: E1204 13:58:48.181503 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.209914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.209959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.209989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.210009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.210021 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.312181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.312276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.312293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.312318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.312335 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.415531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.415571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.415580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.415595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.415607 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.517635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.517669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.517677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.517690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.517699 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.620451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.620505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.620516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.620533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.620545 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.722643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.722681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.722690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.722707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.722717 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.825580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.825643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.825675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.825705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.825728 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.928448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.928493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.928504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.928520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:48 crc kubenswrapper[4715]: I1204 13:58:48.928529 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:48Z","lastTransitionTime":"2025-12-04T13:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.031162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.031358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.031387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.031418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.031442 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.133325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.133373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.133384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.133402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.133414 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.236358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.236417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.236432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.236454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.236470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.339289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.339330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.339341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.339356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.339367 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.441642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.441710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.441726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.441749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.441769 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.544001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.544074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.544093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.544114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.544126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.646920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.646974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.646985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.647001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.647011 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.749733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.749784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.749797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.749820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.749834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.852295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.852334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.852345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.852360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.852370 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.954350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.954403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.954415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.954432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:49 crc kubenswrapper[4715]: I1204 13:58:49.954443 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:49Z","lastTransitionTime":"2025-12-04T13:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.057972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.058021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.058063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.058083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.058097 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.159994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.160070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.160081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.160097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.160111 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.180525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.180567 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.180627 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.180684 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.180864 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.180987 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.181113 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.181202 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.262137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.262202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.262215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.262235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.262248 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.364745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.364777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.364786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.364800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.364809 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.466627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.466654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.466661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.466675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.466709 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.569151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.569201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.569214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.569232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.569244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.618848 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.618949 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:59:54.618927184 +0000 UTC m=+151.687645399 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.618971 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.619021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.619148 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.619170 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.619188 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:59:54.6191784 +0000 UTC m=+151.687896615 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.619210 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:59:54.61919682 +0000 UTC m=+151.687915035 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.671711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.671752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.671762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.671779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.671791 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.720218 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.720263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720380 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720424 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720436 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720486 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:59:54.720470697 +0000 UTC m=+151.789188982 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720387 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720521 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720533 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:50 crc kubenswrapper[4715]: E1204 13:58:50.720595 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:59:54.720566669 +0000 UTC m=+151.789284934 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.774129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.774169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.774178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.774219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.774228 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.876001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.876090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.876103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.876118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.876129 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.978442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.978707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.978796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.978869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:50 crc kubenswrapper[4715]: I1204 13:58:50.978935 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:50Z","lastTransitionTime":"2025-12-04T13:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.081314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.081813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.081897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.081972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.082093 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.184242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.184557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.184653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.184782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.184864 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.288113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.288382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.288467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.288559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.288638 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.390680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.390721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.390758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.390774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.390785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.492644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.492691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.492702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.492717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.492729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.595442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.595507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.595517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.595529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.595539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.699399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.699473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.699492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.699516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.699534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.802185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.802239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.802250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.802268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.802279 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.905382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.905432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.905440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.905456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:51 crc kubenswrapper[4715]: I1204 13:58:51.905466 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:51Z","lastTransitionTime":"2025-12-04T13:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.008135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.008190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.008204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.008232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.008249 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.112092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.112145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.112154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.112171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.112181 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.180023 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.180120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:52 crc kubenswrapper[4715]: E1204 13:58:52.180191 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.180212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:52 crc kubenswrapper[4715]: E1204 13:58:52.180301 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.180342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:52 crc kubenswrapper[4715]: E1204 13:58:52.180375 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:52 crc kubenswrapper[4715]: E1204 13:58:52.180456 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.215874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.215939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.215956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.215986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.216008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.318946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.318995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.319005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.319021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.319074 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.421537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.421841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.421918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.421995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.422125 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.524568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.524860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.524951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.525080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.525175 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.627745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.627787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.627795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.627810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.627819 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.729880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.729926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.729935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.729949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.729958 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.833062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.833104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.833117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.833136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.833149 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.935263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.935302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.935313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.935330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:52 crc kubenswrapper[4715]: I1204 13:58:52.935340 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:52Z","lastTransitionTime":"2025-12-04T13:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.037821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.038079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.038213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.038246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.038260 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.140567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.140609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.140620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.140638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.140652 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.196396 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.206609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.227387 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.243446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.243498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.243512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.243533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.243546 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.243980 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.260846 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.270679 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.282149 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.304656 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.317204 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.336700 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.346123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.346184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.346200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.346229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.346247 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.350388 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.363061 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.378558 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.394996 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.409401 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.428518 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.440734 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.448389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.448444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.448461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.448488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.448519 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.456246 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.473899 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.550809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.550857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.550868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.550889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.550901 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.653638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.653678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.653687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.653699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.653708 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.756547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.756636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.756653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.756684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.756702 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.860369 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.860420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.860432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.860450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.860463 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.963592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.963636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.963645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.963660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:53 crc kubenswrapper[4715]: I1204 13:58:53.963671 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:53Z","lastTransitionTime":"2025-12-04T13:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.065688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.065736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.065750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.065767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.065780 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.168705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.168783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.168806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.168838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.168863 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.180076 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.180230 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.180424 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.180442 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.180413 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.180622 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.180717 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.180844 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.271646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.271699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.271713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.271733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.271748 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.374793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.374852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.374861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.374879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.374894 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.477438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.477484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.477499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.477516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.477526 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.580136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.580223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.580251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.580282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.580305 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.682796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.682857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.682867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.682883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.682894 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.694254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.694296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.694307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.694324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.694335 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.706131 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.709385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.709433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.709446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.709468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.709481 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.721790 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.725093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.725146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.725159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.725174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.725185 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.735772 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.738864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.738897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.738905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.738919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.738928 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.748577 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.751710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.751744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.751755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.751770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.751782 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.763055 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:58:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:58:54 crc kubenswrapper[4715]: E1204 13:58:54.763193 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.784415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.784481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.784494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.784509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.784520 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.886787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.886827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.886838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.886857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.886868 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.989662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.989711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.989728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.989747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:54 crc kubenswrapper[4715]: I1204 13:58:54.989757 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:54Z","lastTransitionTime":"2025-12-04T13:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.091775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.091834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.091851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.091876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.091893 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.194170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.194208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.194216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.194231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.194240 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.297610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.297675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.297691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.297717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.297734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.400534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.400828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.400841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.400857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.400873 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.502981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.503010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.503020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.503049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.503059 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.606159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.606630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.606804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.606968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.607175 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.710054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.710301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.710363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.710425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.710544 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.813585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.813625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.813830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.813850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.813866 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.916591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.916955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.917240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.917441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:55 crc kubenswrapper[4715]: I1204 13:58:55.917585 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:55Z","lastTransitionTime":"2025-12-04T13:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.021757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.022185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.022378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.022549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.022717 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.126743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.126803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.126822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.126846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.126865 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.180698 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.180784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.180707 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:56 crc kubenswrapper[4715]: E1204 13:58:56.180910 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:56 crc kubenswrapper[4715]: E1204 13:58:56.181014 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:56 crc kubenswrapper[4715]: E1204 13:58:56.181153 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.180734 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:56 crc kubenswrapper[4715]: E1204 13:58:56.181500 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.230485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.231150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.231186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.231216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.231238 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.334540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.334603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.334621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.334648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.334666 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.436694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.436729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.436757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.436771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.436780 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.541055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.541119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.541131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.541154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.541170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.644569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.644635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.644647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.644670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.644688 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.747728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.748005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.748159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.748316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.748448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.851301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.851362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.851374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.851395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.851407 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.953967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.954004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.954016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.954057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:56 crc kubenswrapper[4715]: I1204 13:58:56.954073 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:56Z","lastTransitionTime":"2025-12-04T13:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.056780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.056833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.056841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.056855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.056865 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.159200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.159284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.159294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.159313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.159323 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.180350 4715 scope.go:117] "RemoveContainer" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" Dec 04 13:58:57 crc kubenswrapper[4715]: E1204 13:58:57.180577 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.261149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.261190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.261201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.261217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.261229 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.363293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.363334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.363343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.363356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.363366 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.466209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.466262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.466274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.466295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.466309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.568757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.569018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.569143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.569229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.569309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.672264 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.672306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.672317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.672335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.672356 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.774669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.774711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.774723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.774739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.774749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.878216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.878273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.878293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.878322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.878343 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.987080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.987160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.987172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.987190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:57 crc kubenswrapper[4715]: I1204 13:58:57.987205 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:57Z","lastTransitionTime":"2025-12-04T13:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.090118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.090163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.090175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.090192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.090205 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.179810 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.179863 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.179833 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:58:58 crc kubenswrapper[4715]: E1204 13:58:58.179973 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.179826 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:58:58 crc kubenswrapper[4715]: E1204 13:58:58.180104 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:58:58 crc kubenswrapper[4715]: E1204 13:58:58.180205 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:58:58 crc kubenswrapper[4715]: E1204 13:58:58.180293 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.192961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.193023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.193102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.193277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.193308 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.295894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.295961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.295979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.296004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.296023 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.399210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.399263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.399285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.399304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.399320 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.502101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.502153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.502167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.502188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.502202 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.604987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.605070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.605086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.605125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.605139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.707343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.707388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.707410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.707431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.707446 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.810630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.810692 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.810702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.810719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.810729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.912462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.912716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.912802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.912895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:58 crc kubenswrapper[4715]: I1204 13:58:58.912980 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:58Z","lastTransitionTime":"2025-12-04T13:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.016266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.016319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.016337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.016360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.016377 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.119384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.119673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.119852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.119991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.120156 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.222374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.222696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.222779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.222882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.222965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.325565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.325632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.325657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.325685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.325702 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.427833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.427880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.427896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.427916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.427931 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.530394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.530441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.530452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.530470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.530485 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.633385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.633431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.633441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.633458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.633469 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.736578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.736627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.736638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.736656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.736688 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.839856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.839903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.839914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.839931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.839943 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.943439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.943505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.943528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.943599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:58:59 crc kubenswrapper[4715]: I1204 13:58:59.943624 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:58:59Z","lastTransitionTime":"2025-12-04T13:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.047771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.047884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.047933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.047974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.048001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.152031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.152527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.152749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.152965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.153003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.179822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.179864 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:00 crc kubenswrapper[4715]: E1204 13:59:00.180519 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.180761 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.180842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:00 crc kubenswrapper[4715]: E1204 13:59:00.180982 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:00 crc kubenswrapper[4715]: E1204 13:59:00.181148 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:00 crc kubenswrapper[4715]: E1204 13:59:00.181323 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.256648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.257220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.257427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.257712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.257901 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.361686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.361754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.361772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.361800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.361818 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.465026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.465117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.465135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.465162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.465180 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.567609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.567666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.567682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.567711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.567727 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.670428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.670475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.670491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.670525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.670540 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.772498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.772528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.772536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.772549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.772558 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.874674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.874733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.874744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.874762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.874775 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.978291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.978333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.978341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.978357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:00 crc kubenswrapper[4715]: I1204 13:59:00.978367 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:00Z","lastTransitionTime":"2025-12-04T13:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.081742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.081823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.081859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.081891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.081909 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.184671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.184720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.184732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.184749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.184760 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.288275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.288334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.288348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.288372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.288388 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.392025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.392128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.392145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.392175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.392197 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.495537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.495618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.495642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.495675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.495698 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.598607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.598665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.598677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.598694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.598705 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.701017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.701079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.701087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.701102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.701110 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.803930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.803970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.803980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.804000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.804014 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.907061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.907100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.907107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.907139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:01 crc kubenswrapper[4715]: I1204 13:59:01.907148 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:01Z","lastTransitionTime":"2025-12-04T13:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.009365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.009425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.009434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.009448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.009458 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.113664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.113722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.113740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.113767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.113785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.180501 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.180586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.180630 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:02 crc kubenswrapper[4715]: E1204 13:59:02.180635 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.180515 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:02 crc kubenswrapper[4715]: E1204 13:59:02.180709 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:02 crc kubenswrapper[4715]: E1204 13:59:02.180852 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:02 crc kubenswrapper[4715]: E1204 13:59:02.180963 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.216477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.216541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.216566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.216604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.216626 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.319422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.319475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.319484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.319500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.319511 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.422485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.422526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.422535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.422553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.422564 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.525150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.525188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.525198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.525213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.525223 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.628051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.628088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.628097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.628111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.628122 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.730630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.730670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.730680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.730695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.730706 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.833609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.833668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.833677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.833692 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.833701 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.936014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.936081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.936093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.936112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:02 crc kubenswrapper[4715]: I1204 13:59:02.936124 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:02Z","lastTransitionTime":"2025-12-04T13:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.040563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.040600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.040609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.040621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.040631 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.143615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.143671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.143681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.143704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.143720 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.197412 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.215517 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.231933 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.243400 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.246909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.246937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.246946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.246978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.246988 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.255753 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.265796 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.282403 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.295861 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.307326 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.324191 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.340891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.349363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.349417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.349433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.349455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.349472 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.356527 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.371695 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.384160 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.397760 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.410664 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.425347 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.438190 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.452118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.452170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.452183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.452202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.452213 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.454288 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:03Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.554670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.554708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.554727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.554746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.554757 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.657090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.657139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.657148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.657162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.657170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.758845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.759274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.759508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.759868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.760215 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.862754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.862792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.862802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.862818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.862828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.965741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.965804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.965826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.965852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:03 crc kubenswrapper[4715]: I1204 13:59:03.965870 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:03Z","lastTransitionTime":"2025-12-04T13:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.068000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.068121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.068147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.068172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.068189 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.171486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.171537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.171552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.171575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.171589 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.180478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.180928 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.180858 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.180810 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.181522 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.181838 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.182138 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.181972 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.274514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.274626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.274646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.274665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.274674 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.376919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.376959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.376972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.376991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.377003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.483614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.483729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.483840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.483871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.483911 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.587615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.587695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.587722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.587757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.587795 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.690439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.690507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.690519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.690539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.690550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.793159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.793196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.793204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.793217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.793226 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.895512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.895564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.895573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.895591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.895603 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.935881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.935922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.935932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.935946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.935956 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.951128 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.956318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.956344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.956353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.956394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.956404 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.977166 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.982266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.982341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.982376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.982419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:04 crc kubenswrapper[4715]: I1204 13:59:04.982446 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:04Z","lastTransitionTime":"2025-12-04T13:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:04 crc kubenswrapper[4715]: E1204 13:59:04.999659 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:04Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.004527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.004766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.004901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.005072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.005208 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: E1204 13:59:05.020797 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.025691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.025762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.025784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.025812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.025832 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: E1204 13:59:05.042470 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:05Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:05 crc kubenswrapper[4715]: E1204 13:59:05.042903 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.045423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.045484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.045501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.045531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.045551 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.148636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.148700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.148721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.148748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.148766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.252017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.252085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.252101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.252122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.252138 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.355187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.355242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.355254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.355276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.355294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.458537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.458623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.458645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.458674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.458691 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.561947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.561990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.562001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.562019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.562061 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.664759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.664815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.664832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.664852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.664868 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.767334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.767374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.767384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.767402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.767413 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.869803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.869842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.869852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.869867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.869877 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.972957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.973013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.973060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.973083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:05 crc kubenswrapper[4715]: I1204 13:59:05.973100 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:05Z","lastTransitionTime":"2025-12-04T13:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.074626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.074657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.074665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.074677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.074686 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.177165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.177200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.177209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.177224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.177235 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.179590 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.179630 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.179712 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:06 crc kubenswrapper[4715]: E1204 13:59:06.179853 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.179974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:06 crc kubenswrapper[4715]: E1204 13:59:06.180117 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:06 crc kubenswrapper[4715]: E1204 13:59:06.180244 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:06 crc kubenswrapper[4715]: E1204 13:59:06.180189 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.279027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.279075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.279083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.279097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.279106 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.381338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.381368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.381376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.381388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.381397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.484200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.484244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.484256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.484272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.484284 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.586864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.586930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.586948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.586975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.586993 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.689319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.689352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.689361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.689377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.689387 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.791200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.791234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.791244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.791259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.791269 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.894501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.894588 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.894614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.894643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.894661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.997696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.997756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.997774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.997836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:06 crc kubenswrapper[4715]: I1204 13:59:06.997853 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:06Z","lastTransitionTime":"2025-12-04T13:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.100699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.100738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.100746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.100761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.100770 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.203106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.203147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.203157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.203173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.203185 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.306204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.306245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.306257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.306275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.306287 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.408737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.408778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.408788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.408805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.408814 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.510558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.510600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.510608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.510623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.510631 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.613634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.613692 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.613707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.613727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.613742 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.716580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.716629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.716646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.716669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.716688 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.818476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.818507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.818516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.818529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.818539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.920985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.921028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.921065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.921086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:07 crc kubenswrapper[4715]: I1204 13:59:07.921098 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:07Z","lastTransitionTime":"2025-12-04T13:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.024275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.024332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.024350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.024378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.024401 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.127170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.127241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.127265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.127292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.127314 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.180664 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:08 crc kubenswrapper[4715]: E1204 13:59:08.180856 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.181324 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.181451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:08 crc kubenswrapper[4715]: E1204 13:59:08.181588 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.181518 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.181938 4715 scope.go:117] "RemoveContainer" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" Dec 04 13:59:08 crc kubenswrapper[4715]: E1204 13:59:08.181966 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:08 crc kubenswrapper[4715]: E1204 13:59:08.182126 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:08 crc kubenswrapper[4715]: E1204 13:59:08.182243 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.230228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.230284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.230301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.230325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.230342 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.333138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.333601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.333959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.334219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.334437 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.437724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.437779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.437798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.437823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.437841 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.539269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.539330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.539345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.539368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.539380 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.642571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.642648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.642674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.642704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.642724 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.745026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.745073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.745082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.745095 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.745103 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.847652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.847744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.847784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.847814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.847837 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.951401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.951448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.951456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.951471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:08 crc kubenswrapper[4715]: I1204 13:59:08.951482 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:08Z","lastTransitionTime":"2025-12-04T13:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.053645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.053689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.053697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.053711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.053720 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.155649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.155731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.155756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.155785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.155807 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.258986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.259125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.259154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.259187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.259207 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.361400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.361443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.361454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.361468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.361482 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.463620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.463672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.463689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.463708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.463718 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.566002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.566078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.566087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.566101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.566109 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.668500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.668531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.668540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.668553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.668561 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.771228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.771295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.771312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.771339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.771359 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.873747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.873803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.873817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.873835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.873846 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.926958 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:09 crc kubenswrapper[4715]: E1204 13:59:09.927355 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:59:09 crc kubenswrapper[4715]: E1204 13:59:09.927475 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 14:00:13.927440549 +0000 UTC m=+170.996158824 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.976822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.976913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.976936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.976973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:09 crc kubenswrapper[4715]: I1204 13:59:09.976998 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:09Z","lastTransitionTime":"2025-12-04T13:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.079430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.079492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.079503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.079520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.079532 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.179852 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.179879 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.179887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.179862 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:10 crc kubenswrapper[4715]: E1204 13:59:10.180091 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:10 crc kubenswrapper[4715]: E1204 13:59:10.179966 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:10 crc kubenswrapper[4715]: E1204 13:59:10.180229 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:10 crc kubenswrapper[4715]: E1204 13:59:10.180324 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.181343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.181368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.181378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.181392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.181404 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.284261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.284303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.284314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.284331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.284344 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.387166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.387201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.387209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.387222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.387231 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.491423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.491501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.491519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.491545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.491563 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.595018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.595103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.595125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.595156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.595181 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.697406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.697471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.697494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.697522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.697543 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.800478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.800520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.800529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.800544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.800554 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.903691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.903766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.903800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.903825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:10 crc kubenswrapper[4715]: I1204 13:59:10.903845 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:10Z","lastTransitionTime":"2025-12-04T13:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.006543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.006639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.006699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.006722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.007547 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.111213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.111271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.111279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.111293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.111320 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.214213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.214284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.214299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.214319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.214386 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.318004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.318121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.318154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.318184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.318205 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.420958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.421017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.421031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.421098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.421110 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.524376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.524463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.524483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.524520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.524538 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.628117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.628178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.628194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.628272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.628291 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.730682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.730743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.730761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.730787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.730804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.832919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.832957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.832966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.832978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.832986 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.936813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.936893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.936915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.936945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:11 crc kubenswrapper[4715]: I1204 13:59:11.936963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:11Z","lastTransitionTime":"2025-12-04T13:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.039575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.039630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.039644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.039662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.039675 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.143161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.143217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.143232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.143251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.143266 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.179960 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.180063 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.179981 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.179955 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:12 crc kubenswrapper[4715]: E1204 13:59:12.180217 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:12 crc kubenswrapper[4715]: E1204 13:59:12.180366 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:12 crc kubenswrapper[4715]: E1204 13:59:12.180536 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:12 crc kubenswrapper[4715]: E1204 13:59:12.180806 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.247305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.247381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.247403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.247427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.247448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.350709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.350776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.350800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.350826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.350843 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.454221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.454270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.454286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.454312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.454389 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.557329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.557397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.557421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.557448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.557469 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.661003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.661050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.661060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.661074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.661083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.763653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.763710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.763723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.763743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.763758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.867338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.867397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.867414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.867437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.867459 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.970283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.970335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.970353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.970380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:12 crc kubenswrapper[4715]: I1204 13:59:12.970397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:12Z","lastTransitionTime":"2025-12-04T13:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.072521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.072582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.072598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.072620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.072634 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.175227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.175270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.175278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.175292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.175301 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.193670 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.211735 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.224951 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.235985 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.245794 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.265276 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.276416 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.277520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.277552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.277560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.277573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.277585 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.286231 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.299645 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.314577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.333489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.348054 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.360384 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.379915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.379981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.379999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.380024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.380078 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.383799 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.405854 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.419283 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.438541 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.453891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.469570 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.482843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.482895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.482909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.482927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.482944 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.585567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.585595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.585603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.585616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.585625 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.687143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.687189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.687205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.687225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.687241 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.790015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.790114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.790136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.790165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.790187 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.892937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.892975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.892985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.893002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.893012 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.995874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.995936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.995957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.995982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:13 crc kubenswrapper[4715]: I1204 13:59:13.996002 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:13Z","lastTransitionTime":"2025-12-04T13:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.097951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.097987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.097995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.098009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.098019 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.179993 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.180089 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:14 crc kubenswrapper[4715]: E1204 13:59:14.180265 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.180308 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.180397 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:14 crc kubenswrapper[4715]: E1204 13:59:14.180598 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:14 crc kubenswrapper[4715]: E1204 13:59:14.180814 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:14 crc kubenswrapper[4715]: E1204 13:59:14.180917 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.201011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.201136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.201157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.201186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.201209 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.304372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.304444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.304465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.304493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.304510 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.407424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.407485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.407502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.407525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.407543 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.511319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.511396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.511417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.511442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.511460 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.613422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.613473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.613485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.613503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.613516 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.715524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.715602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.715636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.715660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.715674 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.819160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.819292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.819317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.819351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.819376 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.921861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.921938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.921978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.922015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:14 crc kubenswrapper[4715]: I1204 13:59:14.922086 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:14Z","lastTransitionTime":"2025-12-04T13:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.024025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.024080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.024089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.024104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.024114 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.071165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.071263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.071281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.071306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.071328 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: E1204 13:59:15.092117 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:15Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.097413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.097471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.097491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.097516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.097532 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: E1204 13:59:15.119526 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:15Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.124516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.124921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.124935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.124951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.124961 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: E1204 13:59:15.141534 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:15Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.145438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.145471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.145481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.145501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.145540 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: E1204 13:59:15.156784 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:15Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.159899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.159948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.159960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.159977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.159989 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: E1204 13:59:15.175022 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:15Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:15 crc kubenswrapper[4715]: E1204 13:59:15.175163 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.176596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.176623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.176631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.176647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.176655 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.279734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.279784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.279797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.279815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.279828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.382109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.382159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.382167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.382181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.382190 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.484153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.484195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.484206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.484222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.484234 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.586913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.586942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.586951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.586964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.586972 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.689390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.689423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.689431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.689444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.689455 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.792696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.792743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.792754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.792772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.792785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.895475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.895575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.895596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.895619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.895636 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.997870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.997901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.997910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.997923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:15 crc kubenswrapper[4715]: I1204 13:59:15.997931 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:15Z","lastTransitionTime":"2025-12-04T13:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.101336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.101414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.101427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.101469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.101482 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.180951 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:16 crc kubenswrapper[4715]: E1204 13:59:16.181167 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.181298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:16 crc kubenswrapper[4715]: E1204 13:59:16.181440 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.181453 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.181479 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:16 crc kubenswrapper[4715]: E1204 13:59:16.181528 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:16 crc kubenswrapper[4715]: E1204 13:59:16.181589 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.204621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.204702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.204715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.204743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.204761 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.307810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.307857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.307868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.307885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.307895 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.410792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.410858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.410882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.410912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.410932 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.513802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.513965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.513976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.513991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.514022 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.616407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.616477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.616494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.616520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.616537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.718932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.718977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.718989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.719008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.719020 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.821312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.821362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.821374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.821391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.821403 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.924277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.924331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.924343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.924361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:16 crc kubenswrapper[4715]: I1204 13:59:16.924375 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:16Z","lastTransitionTime":"2025-12-04T13:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.026953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.027002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.027018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.027071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.027088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.129173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.129232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.129249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.129273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.129290 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.231510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.231632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.231647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.231666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.231712 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.334660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.334711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.334721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.334737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.334749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.437416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.437474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.437491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.437513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.437529 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.539874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.539926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.539942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.539964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.539980 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.642223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.642270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.642282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.642303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.642317 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.744530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.744590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.744612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.744638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.744656 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.847744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.847812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.847835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.847869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.847893 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.951925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.951963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.951972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.951985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:17 crc kubenswrapper[4715]: I1204 13:59:17.951994 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:17Z","lastTransitionTime":"2025-12-04T13:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.054958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.054989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.054996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.055011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.055020 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.157252 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.157297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.157305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.157319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.157328 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.180356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.180432 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.180397 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:18 crc kubenswrapper[4715]: E1204 13:59:18.180549 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.180570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:18 crc kubenswrapper[4715]: E1204 13:59:18.180650 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:18 crc kubenswrapper[4715]: E1204 13:59:18.180718 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:18 crc kubenswrapper[4715]: E1204 13:59:18.180791 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.260083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.260131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.260146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.260166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.260180 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.362997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.363060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.363073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.363089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.363100 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.467751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.467799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.467810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.467829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.467845 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.570836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.570904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.570923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.570951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.570969 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.673969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.674018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.674052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.674071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.674082 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.776591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.776678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.776711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.776777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.776801 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.879141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.879184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.879194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.879209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.879244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.982148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.982225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.982247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.982272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:18 crc kubenswrapper[4715]: I1204 13:59:18.982289 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:18Z","lastTransitionTime":"2025-12-04T13:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.085299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.085358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.085377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.085400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.085420 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.187115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.187558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.187662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.187807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.187913 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.290453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.290498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.290522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.290545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.290553 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.392893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.392928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.392936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.392954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.392971 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.495560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.495618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.495635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.495658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.495674 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.597660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.597693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.597701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.597714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.597723 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.700211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.700271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.700310 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.700348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.700370 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.803113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.803189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.803214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.803241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.803260 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.906669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.906724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.906741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.906764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:19 crc kubenswrapper[4715]: I1204 13:59:19.906782 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:19Z","lastTransitionTime":"2025-12-04T13:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.008649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.008703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.008720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.008744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.008761 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:20Z","lastTransitionTime":"2025-12-04T13:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.111001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.111064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.111077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.111123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.111136 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:20Z","lastTransitionTime":"2025-12-04T13:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.180228 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.180309 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.180434 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:20 crc kubenswrapper[4715]: E1204 13:59:20.180420 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.180528 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:20 crc kubenswrapper[4715]: E1204 13:59:20.180649 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:20 crc kubenswrapper[4715]: E1204 13:59:20.180739 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:20 crc kubenswrapper[4715]: E1204 13:59:20.180812 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.213694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.213730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.213737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.213751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.213761 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:20Z","lastTransitionTime":"2025-12-04T13:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.316152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.316189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.316196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.316209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.316221 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:20Z","lastTransitionTime":"2025-12-04T13:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.849122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.849177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.849196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.849223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.849248 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:20Z","lastTransitionTime":"2025-12-04T13:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.951532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.951586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.951599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.951616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:20 crc kubenswrapper[4715]: I1204 13:59:20.951627 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:20Z","lastTransitionTime":"2025-12-04T13:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.054106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.054149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.054178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.054197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.054211 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.156555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.156607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.156625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.156649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.156667 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.182408 4715 scope.go:117] "RemoveContainer" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.258877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.259138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.259146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.259161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.259170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.362527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.362608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.362626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.362685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.362702 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.465338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.465391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.465401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.465416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.465424 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.567685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.567712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.567720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.567734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.567743 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.670450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.670509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.670521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.670539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.670552 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.772753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.772785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.772793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.772806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.772816 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.874917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.874963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.874975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.874993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.875007 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.905136 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/3.log" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.908272 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.909070 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.925714 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:21Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.946408 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:21Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.959723 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:21Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.970985 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:21Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.976665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.976712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.976725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.976744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.976783 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:21Z","lastTransitionTime":"2025-12-04T13:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.983679 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:21Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:21 crc kubenswrapper[4715]: I1204 13:59:21.994320 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:21Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.004241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.014779 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.026796 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.038100 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.052442 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.069695 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:59:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.078188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.078283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.078354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.078416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.078480 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.084100 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.092804 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.108788 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.121485 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.135475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.146181 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.160253 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.179879 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.179963 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.180331 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.180374 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:22 crc kubenswrapper[4715]: E1204 13:59:22.180425 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:22 crc kubenswrapper[4715]: E1204 13:59:22.180504 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:22 crc kubenswrapper[4715]: E1204 13:59:22.180564 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:22 crc kubenswrapper[4715]: E1204 13:59:22.180611 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.181754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.181803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.181847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.181865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.181876 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.284773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.284843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.284855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.284872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.284909 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.387385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.387435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.387445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.387464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.387475 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.490111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.490169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.490185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.490205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.490216 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.592798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.592851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.592866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.592888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.592903 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.694576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.694611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.694623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.694638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.694648 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.796655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.796687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.796697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.796710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.796720 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.899464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.899504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.899514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.899531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.899542 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:22Z","lastTransitionTime":"2025-12-04T13:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.913120 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/4.log" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.914197 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/3.log" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.917479 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" exitCode=1 Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.917516 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.917571 4715 scope.go:117] "RemoveContainer" containerID="f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.918192 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 13:59:22 crc kubenswrapper[4715]: E1204 13:59:22.918338 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.936923 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.962844 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.980572 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:22 crc kubenswrapper[4715]: I1204 13:59:22.993593 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:22Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.001937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.001969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.001981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.001998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.002007 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:23Z","lastTransitionTime":"2025-12-04T13:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.003205 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.012349 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.029515 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:59:22Z\\\",\\\"message\\\":\\\"er 6\\\\nI1204 13:59:22.015557 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:59:22.015570 7040 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:59:22.015576 7040 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:59:22.015887 7040 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:59:22.015944 7040 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:59:22.015975 7040 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:59:22.016406 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:59:22.016433 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:59:22.016460 7040 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:59:22.016483 7040 factory.go:656] Stopping watch factory\\\\nI1204 13:59:22.016494 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:59:22.016510 7040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:59:22.016521 7040 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:59:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.043812 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.059286 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.070744 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.084493 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.095370 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.103720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.103754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.103764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.103781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.103793 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:23Z","lastTransitionTime":"2025-12-04T13:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.108352 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.118156 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.128258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.138972 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.150433 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.164258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.180517 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.192987 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: E1204 13:59:23.204267 4715 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.204707 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.215157 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.225546 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.239744 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.251483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: E1204 13:59:23.261078 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.262605 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.272388 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.281793 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.293928 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.306609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.317336 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.333595 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f975a43e7262cf919b68b065f7f13e1a829d5617c94ab752346023f5cff68b7f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:41Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 13:58:40.908698 6663 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]stri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:59:22Z\\\",\\\"message\\\":\\\"er 6\\\\nI1204 13:59:22.015557 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:59:22.015570 7040 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:59:22.015576 7040 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:59:22.015887 7040 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:59:22.015944 7040 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:59:22.015975 7040 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:59:22.016406 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:59:22.016433 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:59:22.016460 7040 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:59:22.016483 7040 factory.go:656] Stopping watch factory\\\\nI1204 13:59:22.016494 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:59:22.016510 7040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:59:22.016521 7040 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:59:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.346442 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.358257 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.376195 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.388689 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.399415 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.409431 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.920751 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/4.log" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.924451 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 13:59:23 crc kubenswrapper[4715]: E1204 13:59:23.924813 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.937539 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.949977 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.960983 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.970845 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:23 crc kubenswrapper[4715]: I1204 13:59:23.987273 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:59:22Z\\\",\\\"message\\\":\\\"er 6\\\\nI1204 13:59:22.015557 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:59:22.015570 7040 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:59:22.015576 7040 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:59:22.015887 7040 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:59:22.015944 7040 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:59:22.015975 7040 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:59:22.016406 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:59:22.016433 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:59:22.016460 7040 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:59:22.016483 7040 factory.go:656] Stopping watch factory\\\\nI1204 13:59:22.016494 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:59:22.016510 7040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:59:22.016521 7040 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:59:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.000257 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:23Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.011373 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.027396 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.038145 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.049103 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.087056 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.108483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.119780 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.128800 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.137058 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.148305 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.159314 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.172971 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.179658 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.179658 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.179691 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:24 crc kubenswrapper[4715]: E1204 13:59:24.179766 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.179818 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:24 crc kubenswrapper[4715]: E1204 13:59:24.179909 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:24 crc kubenswrapper[4715]: E1204 13:59:24.180051 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:24 crc kubenswrapper[4715]: E1204 13:59:24.180093 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.188270 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.930454 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/1.log" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.930996 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/0.log" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.931074 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" containerID="64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b" exitCode=1 Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.931111 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerDied","Data":"64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b"} Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.931153 4715 scope.go:117] "RemoveContainer" containerID="bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.932985 4715 scope.go:117] "RemoveContainer" containerID="64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b" Dec 04 13:59:24 crc kubenswrapper[4715]: E1204 13:59:24.933469 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8dlbk_openshift-multus(4ae4f7b7-e122-4bf8-b802-54e882e3d6cc)\"" pod="openshift-multus/multus-8dlbk" podUID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.954889 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2270a1b8-6cf5-4cfe-8773-be7792dc87ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a229013ab58e2d0c6f522d332f73ad8ae89b8b368e1e248279f1690b32b44d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb3f17e6d496b28adb2e9503f332a5b2ed623e5a225d13a7fe7f8c3035f00913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77de4f5ce923e45ad69b550598625cab06858d65e491657bf12ac214a16cf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a5651c6e3e8037e51ab6e4d84163dfa139a14b618fe84a2c6640ad58f0cfd02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.969595 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17e18886-d679-4432-adae-3f3e97f29d87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d2c30e9ac410bf8b2a3f6545317d5be19edcb651e8fddd171397b1e267aa11b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce5a357ebe5742ee83c31212fcfb57c688ed34076a5fb25a64fac31218aeee8d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.981846 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d5ee7c9-f82d-4848-b21f-e529d843456e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5c44507416f208f3eafe61e3c34bb2ffbf7751196ef90de744a7530ebe01bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6v2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:24 crc kubenswrapper[4715]: I1204 13:59:24.993497 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5fb91c3-b578-4416-b8fe-26bd2ed8f886\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3c7f971b45aa2379f73477b420f14e90d7307b8a158df1c52818b7dcbe7cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c861ef085387be3b369bd3349b1595b4e9fc80029d49a73ddb4bbc11a407e5a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f7949955e94d3a48b6817d6d31caadd4c6ef0cf9ded3cae138a9a34ca3da6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:24Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.004761 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.020577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8dlbk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce94efb66697d0ace3bc63ee52bb861a612eaead03a3912b86c026207c5971e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:58:38Z\\\",\\\"message\\\":\\\"2025-12-04T13:57:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080\\\\n2025-12-04T13:57:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_968c54c0-c5a9-423d-bbcb-3951b55bc080 to /host/opt/cni/bin/\\\\n2025-12-04T13:57:53Z [verbose] multus-daemon started\\\\n2025-12-04T13:57:53Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:58:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:59:24Z\\\",\\\"message\\\":\\\"2025-12-04T13:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf7fecd0-9b64-4acc-8018-a8a7e347e136\\\\n2025-12-04T13:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf7fecd0-9b64-4acc-8018-a8a7e347e136 to /host/opt/cni/bin/\\\\n2025-12-04T13:58:39Z [verbose] multus-daemon started\\\\n2025-12-04T13:58:39Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6v25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8dlbk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.034165 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e9fc7cfbc2fab9370a9cdbd6fe2f26b2ce75a38cbf1165a0acc05b9d9cddb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.047148 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a754c83-0416-4a03-95c4-59a98329e921\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.061845 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.074462 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2r8q9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c81c781-f5fa-4d45-91ff-88a0c057b03f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95d91f0dfb8849c4952fe16cfcfed6e15bc760bb91aac4cf82fbae9e4ba50be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wr64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2r8q9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.086118 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kb882" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0c5e3e9-6182-4cbe-969f-ded0f25c43e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://975b205ca54c69cfd4d48f98b1b87da170b74189f7ac5c9b21b681928ac93750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tr2tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kb882\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.108172 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"727ec519-5a38-4de0-bc0f-69993f95c5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:59:22Z\\\",\\\"message\\\":\\\"er 6\\\\nI1204 13:59:22.015557 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:59:22.015570 7040 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 13:59:22.015576 7040 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 13:59:22.015887 7040 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 13:59:22.015944 7040 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:59:22.015975 7040 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:59:22.016406 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:59:22.016433 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:59:22.016460 7040 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:59:22.016483 7040 factory.go:656] Stopping watch factory\\\\nI1204 13:59:22.016494 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:59:22.016510 7040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:59:22.016521 7040 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:59:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xxk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nw55d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.123726 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e58aa-e858-46aa-9833-11b2c0bc3c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ef91fd116f65c05cf49ef4170a353b6e53a54e1d5480172fdc87adefdbb1aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42d43b10670a6580837aa7b5600052435faf9e10f2a4d87d73b59965519370b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be7cd15b804183d6be2005b45169ef157d0a6dc7ed41a3058fd33ebcbd125491\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28644654cd5463a687ec2f3587c7ac1a49758276fcc511f98f8a6fc4515efcb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26f067e5106d54c2b6c050fa732300ac8f4340dc1ee7ba0e859dd3b6cb6cc44c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92d8b5946de719662581f0a7d3816d8d3bd53c80475ab61caafe0f20e36b0ba7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f0381569577bed0578415f1275422ff2df123d987e8558bf76f73e98d5462f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k9g4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xqzzc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.133492 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cf2kh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c9ttc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.151126 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed1e047-8fb5-49eb-8075-f6bf7d25d0cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac432ff468a4ba9b2cd37225a6a9b96f47a44c4fca410103b23f15f4d1626bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b22deb226df54528bf2e3792d0af3643716a76c7d9714836f9d92cdfeb0ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b7aebbe38cf7a9d5fcdccb3ead189477f1dc4021ee1421de4a7f1ac0a8d0670\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4386eea5342198fa946ac3580709e3fbb21641c7c9fa9bc87c78171ce1ceeb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76905e1319131fb9f0318d5a884da69521208908e21618d8f34bca22acb6c2a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70a9fe5b950e791ea379e505e2c64925d1a9eaa278a411ca73c8065e993e84a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b809fb3efbb5ba01e60c3a1b8f48ba477066733ecc0a4ffd8519ec68b87b311f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://048c96bc9c8ba995468896a3b464e05e4abe18e6b83dbc6ab3fe11f417832d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:57:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:57:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.163517 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.178435 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1021a615feeb9641db4cc9004af4a2b152930164c2b0b1216366275cd0f0c83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65356b1248bb0d9cea5893da76d004b7c531672ce294e062a4a7f7a95de5f8fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:57:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.190739 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d1a1ca5-196c-44a2-88c7-a4cdccc51c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3870e6aae6fb90fa812f2b7ce52b1da2537c7f620d522ba2223a37d03cf54fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37b6d2f93433532956e12ea683ec080f86131caf3111aa6b80706b375361c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmmdq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:58:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j9sxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.202181 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:57:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:58:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec872f2f76c451937dd60bc9f3305e03130c4bec2bb887d18b122e45e1d138c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:58:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.505659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.505884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.505952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.506078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.506164 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:25Z","lastTransitionTime":"2025-12-04T13:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:25 crc kubenswrapper[4715]: E1204 13:59:25.518949 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.522773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.522802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.522810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.522824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.522835 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:25Z","lastTransitionTime":"2025-12-04T13:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:25 crc kubenswrapper[4715]: E1204 13:59:25.534740 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.539740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.539770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.539783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.539801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.539813 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:25Z","lastTransitionTime":"2025-12-04T13:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:25 crc kubenswrapper[4715]: E1204 13:59:25.553556 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.557077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.557132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.557148 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.557164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.557174 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:25Z","lastTransitionTime":"2025-12-04T13:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:25 crc kubenswrapper[4715]: E1204 13:59:25.568412 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.573400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.573461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.573478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.573499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.573515 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:25Z","lastTransitionTime":"2025-12-04T13:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:25 crc kubenswrapper[4715]: E1204 13:59:25.586186 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ab0a07db-261c-49f3-bead-b10436f442e6\\\",\\\"systemUUID\\\":\\\"2e286a6a-d778-45b8-8b58-73ec4dd721bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:59:25Z is after 2025-08-24T17:21:41Z" Dec 04 13:59:25 crc kubenswrapper[4715]: E1204 13:59:25.586451 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:59:25 crc kubenswrapper[4715]: I1204 13:59:25.945551 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/1.log" Dec 04 13:59:26 crc kubenswrapper[4715]: I1204 13:59:26.179848 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:26 crc kubenswrapper[4715]: I1204 13:59:26.179916 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:26 crc kubenswrapper[4715]: I1204 13:59:26.179945 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:26 crc kubenswrapper[4715]: I1204 13:59:26.179995 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:26 crc kubenswrapper[4715]: E1204 13:59:26.180322 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:26 crc kubenswrapper[4715]: E1204 13:59:26.180433 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:26 crc kubenswrapper[4715]: E1204 13:59:26.180504 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:26 crc kubenswrapper[4715]: E1204 13:59:26.180571 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:28 crc kubenswrapper[4715]: I1204 13:59:28.179840 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:28 crc kubenswrapper[4715]: E1204 13:59:28.179963 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:28 crc kubenswrapper[4715]: I1204 13:59:28.180516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:28 crc kubenswrapper[4715]: I1204 13:59:28.180580 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:28 crc kubenswrapper[4715]: E1204 13:59:28.180642 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:28 crc kubenswrapper[4715]: I1204 13:59:28.180683 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:28 crc kubenswrapper[4715]: E1204 13:59:28.180743 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:28 crc kubenswrapper[4715]: E1204 13:59:28.180907 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:28 crc kubenswrapper[4715]: E1204 13:59:28.262174 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:30 crc kubenswrapper[4715]: I1204 13:59:30.180309 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:30 crc kubenswrapper[4715]: I1204 13:59:30.180368 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:30 crc kubenswrapper[4715]: I1204 13:59:30.180374 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:30 crc kubenswrapper[4715]: I1204 13:59:30.180339 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:30 crc kubenswrapper[4715]: E1204 13:59:30.180445 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:30 crc kubenswrapper[4715]: E1204 13:59:30.180702 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:30 crc kubenswrapper[4715]: E1204 13:59:30.180761 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:30 crc kubenswrapper[4715]: E1204 13:59:30.180906 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:32 crc kubenswrapper[4715]: I1204 13:59:32.179880 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:32 crc kubenswrapper[4715]: I1204 13:59:32.179984 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:32 crc kubenswrapper[4715]: E1204 13:59:32.180060 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:32 crc kubenswrapper[4715]: I1204 13:59:32.180091 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:32 crc kubenswrapper[4715]: E1204 13:59:32.180190 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:32 crc kubenswrapper[4715]: E1204 13:59:32.180410 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:32 crc kubenswrapper[4715]: I1204 13:59:32.179910 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:32 crc kubenswrapper[4715]: E1204 13:59:32.181404 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.210105 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=77.210082397 podStartE2EDuration="1m17.210082397s" podCreationTimestamp="2025-12-04 13:58:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.199896879 +0000 UTC m=+130.268615104" watchObservedRunningTime="2025-12-04 13:59:33.210082397 +0000 UTC m=+130.278800612" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.221805 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=50.221786352 podStartE2EDuration="50.221786352s" podCreationTimestamp="2025-12-04 13:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.210380994 +0000 UTC m=+130.279099219" watchObservedRunningTime="2025-12-04 13:59:33.221786352 +0000 UTC m=+130.290504577" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.234776 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podStartSLOduration=102.234755838 podStartE2EDuration="1m42.234755838s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.22170972 +0000 UTC m=+130.290427965" watchObservedRunningTime="2025-12-04 13:59:33.234755838 +0000 UTC m=+130.303474053" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.245887 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=107.245864888 podStartE2EDuration="1m47.245864888s" podCreationTimestamp="2025-12-04 13:57:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.235083806 +0000 UTC m=+130.303802041" watchObservedRunningTime="2025-12-04 13:59:33.245864888 +0000 UTC m=+130.314583103" Dec 04 13:59:33 crc kubenswrapper[4715]: E1204 13:59:33.262703 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.305269 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=109.305251755 podStartE2EDuration="1m49.305251755s" podCreationTimestamp="2025-12-04 13:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.293752795 +0000 UTC m=+130.362471020" watchObservedRunningTime="2025-12-04 13:59:33.305251755 +0000 UTC m=+130.373969970" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.326310 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2r8q9" podStartSLOduration=102.326291877 podStartE2EDuration="1m42.326291877s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.316753725 +0000 UTC m=+130.385471940" watchObservedRunningTime="2025-12-04 13:59:33.326291877 +0000 UTC m=+130.395010092" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.326405 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-kb882" podStartSLOduration=102.3264002 podStartE2EDuration="1m42.3264002s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.326180924 +0000 UTC m=+130.394899139" watchObservedRunningTime="2025-12-04 13:59:33.3264002 +0000 UTC m=+130.395118415" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.367658 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xqzzc" podStartSLOduration=102.367639324 podStartE2EDuration="1m42.367639324s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.366864815 +0000 UTC m=+130.435583050" watchObservedRunningTime="2025-12-04 13:59:33.367639324 +0000 UTC m=+130.436357539" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.399813 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=108.399796667 podStartE2EDuration="1m48.399796667s" podCreationTimestamp="2025-12-04 13:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.39947697 +0000 UTC m=+130.468195185" watchObservedRunningTime="2025-12-04 13:59:33.399796667 +0000 UTC m=+130.468514872" Dec 04 13:59:33 crc kubenswrapper[4715]: I1204 13:59:33.437387 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j9sxz" podStartSLOduration=101.437368212 podStartE2EDuration="1m41.437368212s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:33.436921452 +0000 UTC m=+130.505639677" watchObservedRunningTime="2025-12-04 13:59:33.437368212 +0000 UTC m=+130.506086427" Dec 04 13:59:34 crc kubenswrapper[4715]: I1204 13:59:34.180577 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:34 crc kubenswrapper[4715]: I1204 13:59:34.180659 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:34 crc kubenswrapper[4715]: I1204 13:59:34.180662 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:34 crc kubenswrapper[4715]: I1204 13:59:34.180669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:34 crc kubenswrapper[4715]: E1204 13:59:34.180767 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:34 crc kubenswrapper[4715]: E1204 13:59:34.180866 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:34 crc kubenswrapper[4715]: E1204 13:59:34.180959 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:34 crc kubenswrapper[4715]: E1204 13:59:34.181078 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.181995 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 13:59:35 crc kubenswrapper[4715]: E1204 13:59:35.182354 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.918981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.919492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.919702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.919931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.920172 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:59:35Z","lastTransitionTime":"2025-12-04T13:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.993501 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m"] Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.994254 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.997469 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.997709 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.997934 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 13:59:35 crc kubenswrapper[4715]: I1204 13:59:35.997947 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.110564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9082109a-0b43-4049-a89b-02c17e0937a5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.110617 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9082109a-0b43-4049-a89b-02c17e0937a5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.110642 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9082109a-0b43-4049-a89b-02c17e0937a5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.110680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9082109a-0b43-4049-a89b-02c17e0937a5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.110700 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9082109a-0b43-4049-a89b-02c17e0937a5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.179739 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.179831 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.179758 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.179739 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:36 crc kubenswrapper[4715]: E1204 13:59:36.179906 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:36 crc kubenswrapper[4715]: E1204 13:59:36.179980 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:36 crc kubenswrapper[4715]: E1204 13:59:36.180130 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:36 crc kubenswrapper[4715]: E1204 13:59:36.180200 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9082109a-0b43-4049-a89b-02c17e0937a5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211528 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9082109a-0b43-4049-a89b-02c17e0937a5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9082109a-0b43-4049-a89b-02c17e0937a5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211695 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9082109a-0b43-4049-a89b-02c17e0937a5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211890 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9082109a-0b43-4049-a89b-02c17e0937a5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211938 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9082109a-0b43-4049-a89b-02c17e0937a5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.211953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9082109a-0b43-4049-a89b-02c17e0937a5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.212932 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9082109a-0b43-4049-a89b-02c17e0937a5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.218912 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9082109a-0b43-4049-a89b-02c17e0937a5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.232346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9082109a-0b43-4049-a89b-02c17e0937a5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xzs9m\" (UID: \"9082109a-0b43-4049-a89b-02c17e0937a5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.316763 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.983886 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" event={"ID":"9082109a-0b43-4049-a89b-02c17e0937a5","Type":"ContainerStarted","Data":"2083054ab301a1931df629f9daf81141f9c5d586f3ed3e80f9257275c19ebed0"} Dec 04 13:59:36 crc kubenswrapper[4715]: I1204 13:59:36.984299 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" event={"ID":"9082109a-0b43-4049-a89b-02c17e0937a5","Type":"ContainerStarted","Data":"434c1b449910a663dba848c252a95a8b87e76ff1780c745835a2dd477a1686be"} Dec 04 13:59:37 crc kubenswrapper[4715]: I1204 13:59:37.180612 4715 scope.go:117] "RemoveContainer" containerID="64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b" Dec 04 13:59:37 crc kubenswrapper[4715]: I1204 13:59:37.208996 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xzs9m" podStartSLOduration=106.208978928 podStartE2EDuration="1m46.208978928s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:36.996119272 +0000 UTC m=+134.064837517" watchObservedRunningTime="2025-12-04 13:59:37.208978928 +0000 UTC m=+134.277697143" Dec 04 13:59:37 crc kubenswrapper[4715]: I1204 13:59:37.989075 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/1.log" Dec 04 13:59:37 crc kubenswrapper[4715]: I1204 13:59:37.989142 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerStarted","Data":"25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6"} Dec 04 13:59:38 crc kubenswrapper[4715]: I1204 13:59:38.009337 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8dlbk" podStartSLOduration=107.009319167 podStartE2EDuration="1m47.009319167s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:59:38.009228065 +0000 UTC m=+135.077946290" watchObservedRunningTime="2025-12-04 13:59:38.009319167 +0000 UTC m=+135.078037382" Dec 04 13:59:38 crc kubenswrapper[4715]: I1204 13:59:38.179642 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:38 crc kubenswrapper[4715]: I1204 13:59:38.179708 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:38 crc kubenswrapper[4715]: I1204 13:59:38.179752 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:38 crc kubenswrapper[4715]: I1204 13:59:38.179666 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:38 crc kubenswrapper[4715]: E1204 13:59:38.179772 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:38 crc kubenswrapper[4715]: E1204 13:59:38.179863 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:38 crc kubenswrapper[4715]: E1204 13:59:38.179952 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:38 crc kubenswrapper[4715]: E1204 13:59:38.180027 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:38 crc kubenswrapper[4715]: E1204 13:59:38.264380 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:40 crc kubenswrapper[4715]: I1204 13:59:40.180425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:40 crc kubenswrapper[4715]: I1204 13:59:40.180478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:40 crc kubenswrapper[4715]: I1204 13:59:40.180425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:40 crc kubenswrapper[4715]: E1204 13:59:40.180576 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:40 crc kubenswrapper[4715]: I1204 13:59:40.180558 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:40 crc kubenswrapper[4715]: E1204 13:59:40.180672 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:40 crc kubenswrapper[4715]: E1204 13:59:40.180762 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:40 crc kubenswrapper[4715]: E1204 13:59:40.180858 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:42 crc kubenswrapper[4715]: I1204 13:59:42.179935 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:42 crc kubenswrapper[4715]: I1204 13:59:42.179990 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:42 crc kubenswrapper[4715]: I1204 13:59:42.180128 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:42 crc kubenswrapper[4715]: I1204 13:59:42.180185 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:42 crc kubenswrapper[4715]: E1204 13:59:42.180311 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:42 crc kubenswrapper[4715]: E1204 13:59:42.180402 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:42 crc kubenswrapper[4715]: E1204 13:59:42.180529 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:42 crc kubenswrapper[4715]: E1204 13:59:42.180566 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:43 crc kubenswrapper[4715]: E1204 13:59:43.264953 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:44 crc kubenswrapper[4715]: I1204 13:59:44.180132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:44 crc kubenswrapper[4715]: E1204 13:59:44.180593 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:44 crc kubenswrapper[4715]: I1204 13:59:44.180147 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:44 crc kubenswrapper[4715]: E1204 13:59:44.180690 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:44 crc kubenswrapper[4715]: I1204 13:59:44.180191 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:44 crc kubenswrapper[4715]: E1204 13:59:44.180764 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:44 crc kubenswrapper[4715]: I1204 13:59:44.180143 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:44 crc kubenswrapper[4715]: E1204 13:59:44.180844 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:46 crc kubenswrapper[4715]: I1204 13:59:46.179733 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:46 crc kubenswrapper[4715]: I1204 13:59:46.179845 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:46 crc kubenswrapper[4715]: E1204 13:59:46.179869 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:46 crc kubenswrapper[4715]: I1204 13:59:46.179887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:46 crc kubenswrapper[4715]: I1204 13:59:46.179847 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:46 crc kubenswrapper[4715]: E1204 13:59:46.179973 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:46 crc kubenswrapper[4715]: E1204 13:59:46.180100 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:46 crc kubenswrapper[4715]: E1204 13:59:46.180154 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:47 crc kubenswrapper[4715]: I1204 13:59:47.181172 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 13:59:47 crc kubenswrapper[4715]: E1204 13:59:47.182471 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 13:59:48 crc kubenswrapper[4715]: I1204 13:59:48.180542 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:48 crc kubenswrapper[4715]: E1204 13:59:48.180676 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:48 crc kubenswrapper[4715]: I1204 13:59:48.180542 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:48 crc kubenswrapper[4715]: I1204 13:59:48.180561 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:48 crc kubenswrapper[4715]: I1204 13:59:48.180566 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:48 crc kubenswrapper[4715]: E1204 13:59:48.180756 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:48 crc kubenswrapper[4715]: E1204 13:59:48.180808 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:48 crc kubenswrapper[4715]: E1204 13:59:48.180870 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:48 crc kubenswrapper[4715]: E1204 13:59:48.266552 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:50 crc kubenswrapper[4715]: I1204 13:59:50.180212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:50 crc kubenswrapper[4715]: I1204 13:59:50.180267 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:50 crc kubenswrapper[4715]: I1204 13:59:50.180212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:50 crc kubenswrapper[4715]: E1204 13:59:50.180359 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:50 crc kubenswrapper[4715]: I1204 13:59:50.180486 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:50 crc kubenswrapper[4715]: E1204 13:59:50.180676 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:50 crc kubenswrapper[4715]: E1204 13:59:50.180696 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:50 crc kubenswrapper[4715]: E1204 13:59:50.180859 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:52 crc kubenswrapper[4715]: I1204 13:59:52.180158 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:52 crc kubenswrapper[4715]: I1204 13:59:52.180200 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:52 crc kubenswrapper[4715]: I1204 13:59:52.180207 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:52 crc kubenswrapper[4715]: E1204 13:59:52.180291 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:52 crc kubenswrapper[4715]: I1204 13:59:52.180344 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:52 crc kubenswrapper[4715]: E1204 13:59:52.180394 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:52 crc kubenswrapper[4715]: E1204 13:59:52.180462 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:52 crc kubenswrapper[4715]: E1204 13:59:52.180554 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:53 crc kubenswrapper[4715]: E1204 13:59:53.266996 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.180508 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.180562 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.180557 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.180551 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.180762 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.181556 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.181748 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.181855 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.635653 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.635806 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:56.635788015 +0000 UTC m=+273.704506230 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.635844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.635881 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.635962 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.635991 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.635995 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 14:01:56.635988921 +0000 UTC m=+273.704707136 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.636067 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 14:01:56.636026472 +0000 UTC m=+273.704744687 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.736627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:54 crc kubenswrapper[4715]: I1204 13:59:54.736710 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.736916 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.736919 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.737003 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.737024 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.736941 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.737189 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.737166 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 14:01:56.737133285 +0000 UTC m=+273.805851500 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:59:54 crc kubenswrapper[4715]: E1204 13:59:54.737285 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 14:01:56.737259099 +0000 UTC m=+273.805977494 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:59:56 crc kubenswrapper[4715]: I1204 13:59:56.180235 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:56 crc kubenswrapper[4715]: I1204 13:59:56.180356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:56 crc kubenswrapper[4715]: I1204 13:59:56.180260 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:56 crc kubenswrapper[4715]: E1204 13:59:56.180448 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:56 crc kubenswrapper[4715]: E1204 13:59:56.180536 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:56 crc kubenswrapper[4715]: E1204 13:59:56.180679 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:56 crc kubenswrapper[4715]: I1204 13:59:56.180756 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:56 crc kubenswrapper[4715]: E1204 13:59:56.180853 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:58 crc kubenswrapper[4715]: I1204 13:59:58.179582 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:59:58 crc kubenswrapper[4715]: I1204 13:59:58.179615 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:59:58 crc kubenswrapper[4715]: I1204 13:59:58.179649 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:59:58 crc kubenswrapper[4715]: I1204 13:59:58.179717 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 13:59:58 crc kubenswrapper[4715]: E1204 13:59:58.183362 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 13:59:58 crc kubenswrapper[4715]: E1204 13:59:58.183801 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:59:58 crc kubenswrapper[4715]: E1204 13:59:58.184060 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:59:58 crc kubenswrapper[4715]: E1204 13:59:58.184321 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:59:58 crc kubenswrapper[4715]: E1204 13:59:58.268786 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:00 crc kubenswrapper[4715]: I1204 14:00:00.179650 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:00 crc kubenswrapper[4715]: I1204 14:00:00.179685 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:00 crc kubenswrapper[4715]: I1204 14:00:00.179689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:00 crc kubenswrapper[4715]: E1204 14:00:00.179769 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:00 crc kubenswrapper[4715]: I1204 14:00:00.179650 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:00 crc kubenswrapper[4715]: E1204 14:00:00.179865 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:00 crc kubenswrapper[4715]: E1204 14:00:00.179956 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:00 crc kubenswrapper[4715]: E1204 14:00:00.180126 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:02 crc kubenswrapper[4715]: I1204 14:00:02.180381 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:02 crc kubenswrapper[4715]: I1204 14:00:02.180410 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:02 crc kubenswrapper[4715]: E1204 14:00:02.180499 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:02 crc kubenswrapper[4715]: I1204 14:00:02.180394 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:02 crc kubenswrapper[4715]: I1204 14:00:02.180560 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:02 crc kubenswrapper[4715]: E1204 14:00:02.180715 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:02 crc kubenswrapper[4715]: E1204 14:00:02.180770 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:02 crc kubenswrapper[4715]: E1204 14:00:02.180831 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:02 crc kubenswrapper[4715]: I1204 14:00:02.181634 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:00:02 crc kubenswrapper[4715]: E1204 14:00:02.181810 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 14:00:03 crc kubenswrapper[4715]: E1204 14:00:03.269259 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:04 crc kubenswrapper[4715]: I1204 14:00:04.180449 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:04 crc kubenswrapper[4715]: I1204 14:00:04.180590 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:04 crc kubenswrapper[4715]: E1204 14:00:04.180586 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:04 crc kubenswrapper[4715]: I1204 14:00:04.180647 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:04 crc kubenswrapper[4715]: E1204 14:00:04.180722 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:04 crc kubenswrapper[4715]: I1204 14:00:04.180832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:04 crc kubenswrapper[4715]: E1204 14:00:04.180932 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:04 crc kubenswrapper[4715]: E1204 14:00:04.181114 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:06 crc kubenswrapper[4715]: I1204 14:00:06.179785 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:06 crc kubenswrapper[4715]: E1204 14:00:06.180027 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:06 crc kubenswrapper[4715]: I1204 14:00:06.180290 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:06 crc kubenswrapper[4715]: I1204 14:00:06.180334 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:06 crc kubenswrapper[4715]: I1204 14:00:06.180290 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:06 crc kubenswrapper[4715]: E1204 14:00:06.180461 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:06 crc kubenswrapper[4715]: E1204 14:00:06.180617 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:06 crc kubenswrapper[4715]: E1204 14:00:06.180763 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:08 crc kubenswrapper[4715]: I1204 14:00:08.180452 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:08 crc kubenswrapper[4715]: I1204 14:00:08.180483 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:08 crc kubenswrapper[4715]: I1204 14:00:08.180463 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:08 crc kubenswrapper[4715]: I1204 14:00:08.181467 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:08 crc kubenswrapper[4715]: E1204 14:00:08.181599 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:08 crc kubenswrapper[4715]: E1204 14:00:08.181639 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:08 crc kubenswrapper[4715]: E1204 14:00:08.181723 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:08 crc kubenswrapper[4715]: E1204 14:00:08.181803 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:08 crc kubenswrapper[4715]: E1204 14:00:08.271052 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:08 crc kubenswrapper[4715]: I1204 14:00:08.758809 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:00:08 crc kubenswrapper[4715]: I1204 14:00:08.758923 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:00:10 crc kubenswrapper[4715]: I1204 14:00:10.180720 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:10 crc kubenswrapper[4715]: I1204 14:00:10.180719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:10 crc kubenswrapper[4715]: I1204 14:00:10.180849 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:10 crc kubenswrapper[4715]: I1204 14:00:10.181132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:10 crc kubenswrapper[4715]: E1204 14:00:10.181126 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:10 crc kubenswrapper[4715]: E1204 14:00:10.181295 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:10 crc kubenswrapper[4715]: E1204 14:00:10.181386 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:10 crc kubenswrapper[4715]: E1204 14:00:10.181702 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:12 crc kubenswrapper[4715]: I1204 14:00:12.179724 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:12 crc kubenswrapper[4715]: I1204 14:00:12.179789 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:12 crc kubenswrapper[4715]: E1204 14:00:12.179862 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:12 crc kubenswrapper[4715]: I1204 14:00:12.179721 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:12 crc kubenswrapper[4715]: E1204 14:00:12.179923 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:12 crc kubenswrapper[4715]: E1204 14:00:12.179946 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:12 crc kubenswrapper[4715]: I1204 14:00:12.179789 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:12 crc kubenswrapper[4715]: E1204 14:00:12.180012 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:13 crc kubenswrapper[4715]: E1204 14:00:13.272628 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:13 crc kubenswrapper[4715]: I1204 14:00:13.954646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:13 crc kubenswrapper[4715]: E1204 14:00:13.954796 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 14:00:13 crc kubenswrapper[4715]: E1204 14:00:13.954855 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs podName:4cf68f32-a8c0-41c4-a53d-5f3dbd01197e nodeName:}" failed. No retries permitted until 2025-12-04 14:02:15.954841586 +0000 UTC m=+293.023559791 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs") pod "network-metrics-daemon-c9ttc" (UID: "4cf68f32-a8c0-41c4-a53d-5f3dbd01197e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 14:00:14 crc kubenswrapper[4715]: I1204 14:00:14.179859 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:14 crc kubenswrapper[4715]: I1204 14:00:14.179954 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:14 crc kubenswrapper[4715]: E1204 14:00:14.179976 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:14 crc kubenswrapper[4715]: I1204 14:00:14.180021 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:14 crc kubenswrapper[4715]: I1204 14:00:14.180115 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:14 crc kubenswrapper[4715]: E1204 14:00:14.180293 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:14 crc kubenswrapper[4715]: E1204 14:00:14.180418 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:14 crc kubenswrapper[4715]: E1204 14:00:14.180539 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:15 crc kubenswrapper[4715]: I1204 14:00:15.180482 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:00:15 crc kubenswrapper[4715]: E1204 14:00:15.180640 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 14:00:16 crc kubenswrapper[4715]: I1204 14:00:16.180308 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:16 crc kubenswrapper[4715]: I1204 14:00:16.180344 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:16 crc kubenswrapper[4715]: I1204 14:00:16.180344 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:16 crc kubenswrapper[4715]: E1204 14:00:16.180860 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:16 crc kubenswrapper[4715]: E1204 14:00:16.180736 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:16 crc kubenswrapper[4715]: I1204 14:00:16.180386 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:16 crc kubenswrapper[4715]: E1204 14:00:16.180940 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:16 crc kubenswrapper[4715]: E1204 14:00:16.181023 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:18 crc kubenswrapper[4715]: I1204 14:00:18.180401 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:18 crc kubenswrapper[4715]: I1204 14:00:18.180459 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:18 crc kubenswrapper[4715]: E1204 14:00:18.180532 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:18 crc kubenswrapper[4715]: E1204 14:00:18.180598 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:18 crc kubenswrapper[4715]: I1204 14:00:18.180748 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:18 crc kubenswrapper[4715]: I1204 14:00:18.181257 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:18 crc kubenswrapper[4715]: E1204 14:00:18.181371 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:18 crc kubenswrapper[4715]: E1204 14:00:18.181535 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:18 crc kubenswrapper[4715]: E1204 14:00:18.273762 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:20 crc kubenswrapper[4715]: I1204 14:00:20.179921 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:20 crc kubenswrapper[4715]: I1204 14:00:20.179988 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:20 crc kubenswrapper[4715]: I1204 14:00:20.180113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:20 crc kubenswrapper[4715]: E1204 14:00:20.180105 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:20 crc kubenswrapper[4715]: I1204 14:00:20.179933 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:20 crc kubenswrapper[4715]: E1204 14:00:20.180317 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:20 crc kubenswrapper[4715]: E1204 14:00:20.180361 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:20 crc kubenswrapper[4715]: E1204 14:00:20.180518 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:22 crc kubenswrapper[4715]: I1204 14:00:22.180511 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:22 crc kubenswrapper[4715]: I1204 14:00:22.180552 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:22 crc kubenswrapper[4715]: E1204 14:00:22.180761 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:22 crc kubenswrapper[4715]: I1204 14:00:22.180674 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:22 crc kubenswrapper[4715]: E1204 14:00:22.180834 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:22 crc kubenswrapper[4715]: I1204 14:00:22.180552 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:22 crc kubenswrapper[4715]: E1204 14:00:22.180875 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:22 crc kubenswrapper[4715]: E1204 14:00:22.181008 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:23 crc kubenswrapper[4715]: I1204 14:00:23.131449 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/2.log" Dec 04 14:00:23 crc kubenswrapper[4715]: I1204 14:00:23.132501 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/1.log" Dec 04 14:00:23 crc kubenswrapper[4715]: I1204 14:00:23.132648 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" containerID="25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6" exitCode=1 Dec 04 14:00:23 crc kubenswrapper[4715]: I1204 14:00:23.132736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerDied","Data":"25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6"} Dec 04 14:00:23 crc kubenswrapper[4715]: I1204 14:00:23.132835 4715 scope.go:117] "RemoveContainer" containerID="64cdfce02240df93d38072dc69f645cc1dad75492426dc1ca0526b52149c394b" Dec 04 14:00:23 crc kubenswrapper[4715]: I1204 14:00:23.133389 4715 scope.go:117] "RemoveContainer" containerID="25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6" Dec 04 14:00:23 crc kubenswrapper[4715]: E1204 14:00:23.133653 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8dlbk_openshift-multus(4ae4f7b7-e122-4bf8-b802-54e882e3d6cc)\"" pod="openshift-multus/multus-8dlbk" podUID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" Dec 04 14:00:23 crc kubenswrapper[4715]: E1204 14:00:23.274157 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:24 crc kubenswrapper[4715]: I1204 14:00:24.137116 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/2.log" Dec 04 14:00:24 crc kubenswrapper[4715]: I1204 14:00:24.179751 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:24 crc kubenswrapper[4715]: I1204 14:00:24.179811 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:24 crc kubenswrapper[4715]: E1204 14:00:24.179875 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:24 crc kubenswrapper[4715]: E1204 14:00:24.179945 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:24 crc kubenswrapper[4715]: I1204 14:00:24.180065 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:24 crc kubenswrapper[4715]: E1204 14:00:24.180119 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:24 crc kubenswrapper[4715]: I1204 14:00:24.180329 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:24 crc kubenswrapper[4715]: E1204 14:00:24.180415 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:26 crc kubenswrapper[4715]: I1204 14:00:26.180368 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:26 crc kubenswrapper[4715]: I1204 14:00:26.180458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:26 crc kubenswrapper[4715]: E1204 14:00:26.180500 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:26 crc kubenswrapper[4715]: I1204 14:00:26.180565 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:26 crc kubenswrapper[4715]: I1204 14:00:26.180580 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:26 crc kubenswrapper[4715]: E1204 14:00:26.180670 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:26 crc kubenswrapper[4715]: E1204 14:00:26.180750 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:26 crc kubenswrapper[4715]: E1204 14:00:26.180812 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:26 crc kubenswrapper[4715]: I1204 14:00:26.181572 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:00:26 crc kubenswrapper[4715]: E1204 14:00:26.181741 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 14:00:28 crc kubenswrapper[4715]: I1204 14:00:28.179907 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:28 crc kubenswrapper[4715]: I1204 14:00:28.180014 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:28 crc kubenswrapper[4715]: I1204 14:00:28.179907 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:28 crc kubenswrapper[4715]: E1204 14:00:28.180297 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:28 crc kubenswrapper[4715]: E1204 14:00:28.180365 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:28 crc kubenswrapper[4715]: E1204 14:00:28.180106 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:28 crc kubenswrapper[4715]: I1204 14:00:28.180618 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:28 crc kubenswrapper[4715]: E1204 14:00:28.180709 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:28 crc kubenswrapper[4715]: E1204 14:00:28.276164 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:30 crc kubenswrapper[4715]: I1204 14:00:30.179812 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:30 crc kubenswrapper[4715]: I1204 14:00:30.179883 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:30 crc kubenswrapper[4715]: I1204 14:00:30.179912 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:30 crc kubenswrapper[4715]: I1204 14:00:30.179825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:30 crc kubenswrapper[4715]: E1204 14:00:30.179998 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:30 crc kubenswrapper[4715]: E1204 14:00:30.180183 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:30 crc kubenswrapper[4715]: E1204 14:00:30.180276 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:30 crc kubenswrapper[4715]: E1204 14:00:30.180362 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:32 crc kubenswrapper[4715]: I1204 14:00:32.180444 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:32 crc kubenswrapper[4715]: I1204 14:00:32.180463 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:32 crc kubenswrapper[4715]: I1204 14:00:32.180474 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:32 crc kubenswrapper[4715]: I1204 14:00:32.180521 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:32 crc kubenswrapper[4715]: E1204 14:00:32.180817 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:32 crc kubenswrapper[4715]: E1204 14:00:32.181179 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:32 crc kubenswrapper[4715]: E1204 14:00:32.181484 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:32 crc kubenswrapper[4715]: E1204 14:00:32.181772 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:33 crc kubenswrapper[4715]: E1204 14:00:33.276672 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:34 crc kubenswrapper[4715]: I1204 14:00:34.179827 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:34 crc kubenswrapper[4715]: I1204 14:00:34.179862 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:34 crc kubenswrapper[4715]: E1204 14:00:34.179958 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:34 crc kubenswrapper[4715]: I1204 14:00:34.179990 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:34 crc kubenswrapper[4715]: I1204 14:00:34.179999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:34 crc kubenswrapper[4715]: E1204 14:00:34.180121 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:34 crc kubenswrapper[4715]: E1204 14:00:34.180165 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:34 crc kubenswrapper[4715]: E1204 14:00:34.180219 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:36 crc kubenswrapper[4715]: I1204 14:00:36.180011 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:36 crc kubenswrapper[4715]: E1204 14:00:36.180498 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:36 crc kubenswrapper[4715]: I1204 14:00:36.180199 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:36 crc kubenswrapper[4715]: I1204 14:00:36.180237 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:36 crc kubenswrapper[4715]: I1204 14:00:36.180091 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:36 crc kubenswrapper[4715]: I1204 14:00:36.180757 4715 scope.go:117] "RemoveContainer" containerID="25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6" Dec 04 14:00:36 crc kubenswrapper[4715]: E1204 14:00:36.180799 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:36 crc kubenswrapper[4715]: E1204 14:00:36.180918 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:36 crc kubenswrapper[4715]: E1204 14:00:36.181079 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8dlbk_openshift-multus(4ae4f7b7-e122-4bf8-b802-54e882e3d6cc)\"" pod="openshift-multus/multus-8dlbk" podUID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" Dec 04 14:00:36 crc kubenswrapper[4715]: E1204 14:00:36.181285 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.179536 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.179551 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.179637 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:38 crc kubenswrapper[4715]: E1204 14:00:38.179804 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.179817 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:38 crc kubenswrapper[4715]: E1204 14:00:38.179914 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:38 crc kubenswrapper[4715]: E1204 14:00:38.179980 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.180519 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:00:38 crc kubenswrapper[4715]: E1204 14:00:38.180639 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ovnkube-controller pod=ovnkube-node-nw55d_openshift-ovn-kubernetes(727ec519-5a38-4de0-bc0f-69993f95c5ce)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" Dec 04 14:00:38 crc kubenswrapper[4715]: E1204 14:00:38.180825 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:38 crc kubenswrapper[4715]: E1204 14:00:38.278158 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.758413 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:00:38 crc kubenswrapper[4715]: I1204 14:00:38.758499 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:00:40 crc kubenswrapper[4715]: I1204 14:00:40.180451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:40 crc kubenswrapper[4715]: I1204 14:00:40.180541 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:40 crc kubenswrapper[4715]: E1204 14:00:40.180608 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:40 crc kubenswrapper[4715]: E1204 14:00:40.180744 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:40 crc kubenswrapper[4715]: I1204 14:00:40.180809 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:40 crc kubenswrapper[4715]: E1204 14:00:40.180945 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:40 crc kubenswrapper[4715]: I1204 14:00:40.181327 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:40 crc kubenswrapper[4715]: E1204 14:00:40.181427 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:42 crc kubenswrapper[4715]: I1204 14:00:42.179922 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:42 crc kubenswrapper[4715]: I1204 14:00:42.180203 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:42 crc kubenswrapper[4715]: I1204 14:00:42.180211 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:42 crc kubenswrapper[4715]: E1204 14:00:42.180271 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:42 crc kubenswrapper[4715]: E1204 14:00:42.180332 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:42 crc kubenswrapper[4715]: E1204 14:00:42.180394 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:42 crc kubenswrapper[4715]: I1204 14:00:42.179955 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:42 crc kubenswrapper[4715]: E1204 14:00:42.180927 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:43 crc kubenswrapper[4715]: E1204 14:00:43.279301 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:44 crc kubenswrapper[4715]: I1204 14:00:44.180421 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:44 crc kubenswrapper[4715]: I1204 14:00:44.180479 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:44 crc kubenswrapper[4715]: E1204 14:00:44.180568 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:44 crc kubenswrapper[4715]: I1204 14:00:44.180584 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:44 crc kubenswrapper[4715]: I1204 14:00:44.180665 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:44 crc kubenswrapper[4715]: E1204 14:00:44.180738 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:44 crc kubenswrapper[4715]: E1204 14:00:44.180870 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:44 crc kubenswrapper[4715]: E1204 14:00:44.180915 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:46 crc kubenswrapper[4715]: I1204 14:00:46.179877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:46 crc kubenswrapper[4715]: I1204 14:00:46.179993 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:46 crc kubenswrapper[4715]: E1204 14:00:46.180182 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:46 crc kubenswrapper[4715]: I1204 14:00:46.180222 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:46 crc kubenswrapper[4715]: I1204 14:00:46.180203 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:46 crc kubenswrapper[4715]: E1204 14:00:46.180464 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:46 crc kubenswrapper[4715]: E1204 14:00:46.180623 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:46 crc kubenswrapper[4715]: E1204 14:00:46.180719 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:48 crc kubenswrapper[4715]: I1204 14:00:48.180623 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:48 crc kubenswrapper[4715]: E1204 14:00:48.180763 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:48 crc kubenswrapper[4715]: I1204 14:00:48.180841 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:48 crc kubenswrapper[4715]: I1204 14:00:48.180893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:48 crc kubenswrapper[4715]: E1204 14:00:48.180998 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:48 crc kubenswrapper[4715]: E1204 14:00:48.181155 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:48 crc kubenswrapper[4715]: I1204 14:00:48.181354 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:48 crc kubenswrapper[4715]: E1204 14:00:48.181447 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:48 crc kubenswrapper[4715]: E1204 14:00:48.281508 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.180213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.180242 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:50 crc kubenswrapper[4715]: E1204 14:00:50.180335 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.180375 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.180700 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:50 crc kubenswrapper[4715]: E1204 14:00:50.180878 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:50 crc kubenswrapper[4715]: E1204 14:00:50.180972 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.181014 4715 scope.go:117] "RemoveContainer" containerID="25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6" Dec 04 14:00:50 crc kubenswrapper[4715]: E1204 14:00:50.181026 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.181081 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.974370 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-c9ttc"] Dec 04 14:00:50 crc kubenswrapper[4715]: I1204 14:00:50.974471 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:50 crc kubenswrapper[4715]: E1204 14:00:50.974583 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:51 crc kubenswrapper[4715]: I1204 14:00:51.218458 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/2.log" Dec 04 14:00:51 crc kubenswrapper[4715]: I1204 14:00:51.218803 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerStarted","Data":"d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1"} Dec 04 14:00:51 crc kubenswrapper[4715]: I1204 14:00:51.221478 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/4.log" Dec 04 14:00:51 crc kubenswrapper[4715]: I1204 14:00:51.225852 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerStarted","Data":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} Dec 04 14:00:51 crc kubenswrapper[4715]: I1204 14:00:51.226483 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 14:00:51 crc kubenswrapper[4715]: I1204 14:00:51.265559 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podStartSLOduration=180.265538146 podStartE2EDuration="3m0.265538146s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:00:51.265392222 +0000 UTC m=+208.334110467" watchObservedRunningTime="2025-12-04 14:00:51.265538146 +0000 UTC m=+208.334256371" Dec 04 14:00:52 crc kubenswrapper[4715]: I1204 14:00:52.179935 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:52 crc kubenswrapper[4715]: E1204 14:00:52.180075 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 14:00:52 crc kubenswrapper[4715]: I1204 14:00:52.180212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:52 crc kubenswrapper[4715]: I1204 14:00:52.180263 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:52 crc kubenswrapper[4715]: I1204 14:00:52.180212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:52 crc kubenswrapper[4715]: E1204 14:00:52.180327 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 14:00:52 crc kubenswrapper[4715]: E1204 14:00:52.180494 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c9ttc" podUID="4cf68f32-a8c0-41c4-a53d-5f3dbd01197e" Dec 04 14:00:52 crc kubenswrapper[4715]: E1204 14:00:52.180536 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.179629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.179687 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.179633 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.180279 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.181797 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.181882 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.182232 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.183300 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.183524 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 14:00:54 crc kubenswrapper[4715]: I1204 14:00:54.185609 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.381672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.412947 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qzhn8"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.413680 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.415103 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.415522 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.415922 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hb2z4"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.416458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.416563 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.419234 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.419330 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c8mr7"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.419399 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.419703 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.419844 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.419889 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420008 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420085 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420144 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420233 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420243 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420375 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420414 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420512 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420591 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.420748 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.421763 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.426688 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.426750 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.426863 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.427464 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wmltk"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.427568 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.427795 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.428167 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qtq4r"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.428557 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.428852 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.429123 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.429301 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vvwbg"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.429672 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.431227 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.431619 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.431895 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.432194 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.445163 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.445507 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.445567 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.445833 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.451004 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.467083 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.467340 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.467479 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.467596 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.469148 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.469322 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.469504 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.469597 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.470079 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.470225 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.470471 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.470604 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471320 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471391 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471506 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471685 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471819 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471948 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.471960 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472049 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472097 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472132 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472190 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472222 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472274 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472328 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472234 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472466 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472663 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.472873 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.473297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.477311 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nrpl2"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.477731 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.480559 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.480912 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.482432 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.483058 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.483289 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.483445 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.483570 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.485141 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.485484 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.485618 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.486228 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.486468 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.486786 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.486891 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.486997 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.487131 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.489595 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vrf24"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.503901 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hb2z4"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.495476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.503976 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2hpb\" (UniqueName: \"kubernetes.io/projected/96c2ac59-2838-44ca-b4d9-e75092c16301-kube-api-access-h2hpb\") pod \"cluster-samples-operator-665b6dd947-fk5j5\" (UID: \"96c2ac59-2838-44ca-b4d9-e75092c16301\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.503999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-config\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504018 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7b1d7a9-482e-4058-8a97-ae478860ee39-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504050 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-images\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504066 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-config\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504081 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-config\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504096 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-etcd-serving-ca\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-client-ca\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504132 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504160 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-trusted-ca-bundle\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e345625-b45e-479d-9bcb-ca495970b31b-serving-cert\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504204 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdckm\" (UniqueName: \"kubernetes.io/projected/ece56c58-71e4-421a-8dc0-04a543033972-kube-api-access-rdckm\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504218 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgvbc\" (UniqueName: \"kubernetes.io/projected/5e345625-b45e-479d-9bcb-ca495970b31b-kube-api-access-sgvbc\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56596b16-b212-4e17-92ae-0050b15a5382-audit-dir\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504251 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7b1d7a9-482e-4058-8a97-ae478860ee39-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504275 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece56c58-71e4-421a-8dc0-04a543033972-serving-cert\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504291 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m6p5\" (UniqueName: \"kubernetes.io/projected/56596b16-b212-4e17-92ae-0050b15a5382-kube-api-access-9m6p5\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1769830-312d-4f25-870a-b8aaff8b9d0d-config\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504320 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j45m4\" (UniqueName: \"kubernetes.io/projected/e7b1d7a9-482e-4058-8a97-ae478860ee39-kube-api-access-j45m4\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504335 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-config\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs56r\" (UniqueName: \"kubernetes.io/projected/f68f6c4b-010e-47f0-9a1f-74facb7f8190-kube-api-access-bs56r\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504366 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7z2k\" (UniqueName: \"kubernetes.io/projected/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-kube-api-access-m7z2k\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504401 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-oauth-config\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.492818 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504550 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-encryption-config\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504548 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504617 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7b1d7a9-482e-4058-8a97-ae478860ee39-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504666 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-image-import-ca\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-audit\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-config\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-client-ca\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504827 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-serving-cert\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504844 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-service-ca\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504864 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-service-ca-bundle\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/56596b16-b212-4e17-92ae-0050b15a5382-node-pullsecrets\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504924 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e345625-b45e-479d-9bcb-ca495970b31b-config\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.504959 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-serving-cert\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505020 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9h49\" (UniqueName: \"kubernetes.io/projected/c3b08994-cb44-48a9-87d1-af33d2b511b6-kube-api-access-n9h49\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b1769830-312d-4f25-870a-b8aaff8b9d0d-machine-approver-tls\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505109 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gnlf\" (UniqueName: \"kubernetes.io/projected/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-kube-api-access-8gnlf\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505135 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-oauth-serving-cert\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505154 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2gq\" (UniqueName: \"kubernetes.io/projected/261ac4be-400a-48b6-98de-ded0afc94441-kube-api-access-ld2gq\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505188 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68f6c4b-010e-47f0-9a1f-74facb7f8190-serving-cert\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505231 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-etcd-client\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505252 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/96c2ac59-2838-44ca-b4d9-e75092c16301-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fk5j5\" (UID: \"96c2ac59-2838-44ca-b4d9-e75092c16301\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505276 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-config\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505295 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/261ac4be-400a-48b6-98de-ded0afc94441-serving-cert\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505362 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e345625-b45e-479d-9bcb-ca495970b31b-trusted-ca\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505387 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1769830-312d-4f25-870a-b8aaff8b9d0d-auth-proxy-config\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.505403 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfwt\" (UniqueName: \"kubernetes.io/projected/b1769830-312d-4f25-870a-b8aaff8b9d0d-kube-api-access-hvfwt\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.511126 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qzhn8"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.512004 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.513867 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.514017 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.514197 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.515961 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.516142 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.516253 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.516560 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.516659 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.518594 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.518940 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qtq4r"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.519849 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.522438 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.523917 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.524144 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.524304 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.524460 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.524641 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.524804 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.524918 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.526239 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.526423 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.526626 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.526631 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.526761 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.527023 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.528295 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vvwbg"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.529511 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c8mr7"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.529973 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.530282 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.531025 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.531348 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.531640 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.531880 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.532201 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.532256 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.532310 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.535873 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.559903 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.560217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wmltk"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.561942 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.562202 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.562487 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.562915 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.563982 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.564727 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vrf24"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.564765 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.565486 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.569379 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.570153 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nrpl2"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.570201 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.571189 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.579886 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-79f5b"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.580486 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.585188 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-n6w9b"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.585969 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.586492 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-h4k8p"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.587857 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-h4k8p" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.588005 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.588402 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.592170 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-tplh9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.592849 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.593420 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.593961 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.594550 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.595459 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.598418 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.600296 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.600944 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.601660 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m4qn9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.602436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.602532 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.603021 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.607349 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q9ctx"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.607459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-oauth-config\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615319 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-encryption-config\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615348 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/462c99b5-fcc5-47ae-b803-9c49a676afdc-serving-cert\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615379 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-encryption-config\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615404 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7b1d7a9-482e-4058-8a97-ae478860ee39-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615460 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-image-import-ca\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615512 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-audit\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-config\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-client-ca\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615592 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615615 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615652 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-serving-cert\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615676 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-service-ca\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615725 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-service-ca-bundle\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615746 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/56596b16-b212-4e17-92ae-0050b15a5382-node-pullsecrets\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615768 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e345625-b45e-479d-9bcb-ca495970b31b-config\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615788 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-serving-cert\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-etcd-client\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e61a183-e84b-4e3a-8dfb-285670b4e290-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-policies\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9h49\" (UniqueName: \"kubernetes.io/projected/c3b08994-cb44-48a9-87d1-af33d2b511b6-kube-api-access-n9h49\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615917 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b1769830-312d-4f25-870a-b8aaff8b9d0d-machine-approver-tls\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw26f\" (UniqueName: \"kubernetes.io/projected/7e61a183-e84b-4e3a-8dfb-285670b4e290-kube-api-access-lw26f\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615966 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gnlf\" (UniqueName: \"kubernetes.io/projected/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-kube-api-access-8gnlf\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.615988 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-oauth-serving-cert\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.616009 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2gq\" (UniqueName: \"kubernetes.io/projected/261ac4be-400a-48b6-98de-ded0afc94441-kube-api-access-ld2gq\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.616049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw4lw\" (UniqueName: \"kubernetes.io/projected/33d7a759-31d9-4a9c-a613-30db8dc0f3fe-kube-api-access-vw4lw\") pod \"downloads-7954f5f757-vrf24\" (UID: \"33d7a759-31d9-4a9c-a613-30db8dc0f3fe\") " pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.617985 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68f6c4b-010e-47f0-9a1f-74facb7f8190-serving-cert\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-etcd-client\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618162 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-audit-policies\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618190 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/96c2ac59-2838-44ca-b4d9-e75092c16301-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fk5j5\" (UID: \"96c2ac59-2838-44ca-b4d9-e75092c16301\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618214 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-config\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/261ac4be-400a-48b6-98de-ded0afc94441-serving-cert\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618292 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618321 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e61a183-e84b-4e3a-8dfb-285670b4e290-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618376 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcfbdf4d-1139-4153-b4ed-783e13867f47-audit-dir\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618401 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e345625-b45e-479d-9bcb-ca495970b31b-trusted-ca\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618425 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1769830-312d-4f25-870a-b8aaff8b9d0d-auth-proxy-config\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfwt\" (UniqueName: \"kubernetes.io/projected/b1769830-312d-4f25-870a-b8aaff8b9d0d-kube-api-access-hvfwt\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618475 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618499 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618499 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-oauth-config\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618520 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wj72\" (UniqueName: \"kubernetes.io/projected/3f118cab-ae82-49c9-90fb-e00688b70c5b-kube-api-access-2wj72\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2hpb\" (UniqueName: \"kubernetes.io/projected/96c2ac59-2838-44ca-b4d9-e75092c16301-kube-api-access-h2hpb\") pod \"cluster-samples-operator-665b6dd947-fk5j5\" (UID: \"96c2ac59-2838-44ca-b4d9-e75092c16301\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618602 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-config\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618624 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618650 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7b1d7a9-482e-4058-8a97-ae478860ee39-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618673 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-images\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618693 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-config\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618719 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-config\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-etcd-serving-ca\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618773 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-serving-cert\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618805 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-client-ca\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2xcf\" (UniqueName: \"kubernetes.io/projected/462c99b5-fcc5-47ae-b803-9c49a676afdc-kube-api-access-t2xcf\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-trusted-ca-bundle\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618927 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e345625-b45e-479d-9bcb-ca495970b31b-serving-cert\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618954 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdckm\" (UniqueName: \"kubernetes.io/projected/ece56c58-71e4-421a-8dc0-04a543033972-kube-api-access-rdckm\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgvbc\" (UniqueName: \"kubernetes.io/projected/5e345625-b45e-479d-9bcb-ca495970b31b-kube-api-access-sgvbc\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619078 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/462c99b5-fcc5-47ae-b803-9c49a676afdc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619101 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56596b16-b212-4e17-92ae-0050b15a5382-audit-dir\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619193 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7b1d7a9-482e-4058-8a97-ae478860ee39-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619214 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-dir\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece56c58-71e4-421a-8dc0-04a543033972-serving-cert\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619276 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m6p5\" (UniqueName: \"kubernetes.io/projected/56596b16-b212-4e17-92ae-0050b15a5382-kube-api-access-9m6p5\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1769830-312d-4f25-870a-b8aaff8b9d0d-config\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619325 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z8dp\" (UniqueName: \"kubernetes.io/projected/fcfbdf4d-1139-4153-b4ed-783e13867f47-kube-api-access-8z8dp\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619347 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j45m4\" (UniqueName: \"kubernetes.io/projected/e7b1d7a9-482e-4058-8a97-ae478860ee39-kube-api-access-j45m4\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619393 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-config\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619417 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs56r\" (UniqueName: \"kubernetes.io/projected/f68f6c4b-010e-47f0-9a1f-74facb7f8190-kube-api-access-bs56r\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619439 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.619461 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7z2k\" (UniqueName: \"kubernetes.io/projected/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-kube-api-access-m7z2k\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.620284 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.620315 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.621326 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-image-import-ca\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.621438 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.621837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-audit\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.621841 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-blwbc"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.621932 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.622304 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.623760 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-service-ca-bundle\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.623900 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/56596b16-b212-4e17-92ae-0050b15a5382-node-pullsecrets\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.624405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.625716 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-config\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.626187 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-client-ca\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.626991 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e345625-b45e-479d-9bcb-ca495970b31b-config\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.627226 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.636980 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-client-ca\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.637134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7b1d7a9-482e-4058-8a97-ae478860ee39-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.637637 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece56c58-71e4-421a-8dc0-04a543033972-serving-cert\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.637808 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7b1d7a9-482e-4058-8a97-ae478860ee39-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.637959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-config\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.638318 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-encryption-config\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.618808 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.650873 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.651551 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-service-ca\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.652340 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1769830-312d-4f25-870a-b8aaff8b9d0d-config\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.655201 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.660335 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.664656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-config\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.665277 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.665603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/261ac4be-400a-48b6-98de-ded0afc94441-serving-cert\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.667238 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/96c2ac59-2838-44ca-b4d9-e75092c16301-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fk5j5\" (UID: \"96c2ac59-2838-44ca-b4d9-e75092c16301\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.667588 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.667933 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-config\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.667994 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-trusted-ca-bundle\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.668483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.669453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-images\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.669514 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.669866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-etcd-serving-ca\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.670224 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.670443 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68f6c4b-010e-47f0-9a1f-74facb7f8190-config\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.670501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-serving-cert\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.670632 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.670846 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671125 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671331 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671468 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56596b16-b212-4e17-92ae-0050b15a5382-audit-dir\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671350 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671510 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671339 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56596b16-b212-4e17-92ae-0050b15a5382-config\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.671845 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e345625-b45e-479d-9bcb-ca495970b31b-trusted-ca\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673335 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1769830-312d-4f25-870a-b8aaff8b9d0d-auth-proxy-config\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673372 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e345625-b45e-479d-9bcb-ca495970b31b-serving-cert\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56596b16-b212-4e17-92ae-0050b15a5382-etcd-client\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673799 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673804 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-oauth-serving-cert\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-serving-cert\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.673884 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.674469 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b1769830-312d-4f25-870a-b8aaff8b9d0d-machine-approver-tls\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.675007 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.675124 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.675686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.675855 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7jmdt"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.676142 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.676310 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.676681 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rj8j9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.676812 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.677069 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.677239 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.677529 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kppt9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.677749 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.679323 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68f6c4b-010e-47f0-9a1f-74facb7f8190-serving-cert\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.679479 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rfh7t"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.680478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.681782 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683156 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-n6w9b"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683185 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683199 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-79f5b"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683210 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683289 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-h4k8p"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683300 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683312 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m4qn9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683323 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.683422 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.685487 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.686503 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.687805 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.688705 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.689779 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rj8j9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.690801 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q9ctx"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.691807 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.692925 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rfh7t"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.693932 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-blwbc"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.695612 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2lzhm"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.696322 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.696393 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.696477 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.697204 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.698319 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.702010 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.703454 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kppt9"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.704657 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2lzhm"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.705818 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn"] Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.716796 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721142 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-serving-cert\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721176 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2xcf\" (UniqueName: \"kubernetes.io/projected/462c99b5-fcc5-47ae-b803-9c49a676afdc-kube-api-access-t2xcf\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721211 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/462c99b5-fcc5-47ae-b803-9c49a676afdc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721265 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-dir\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721290 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z8dp\" (UniqueName: \"kubernetes.io/projected/fcfbdf4d-1139-4153-b4ed-783e13867f47-kube-api-access-8z8dp\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721306 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721362 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-encryption-config\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721391 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/462c99b5-fcc5-47ae-b803-9c49a676afdc-serving-cert\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721427 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721484 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721501 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e61a183-e84b-4e3a-8dfb-285670b4e290-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.721656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-dir\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.722095 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/462c99b5-fcc5-47ae-b803-9c49a676afdc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.722448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-policies\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.722481 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-etcd-client\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.722498 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw26f\" (UniqueName: \"kubernetes.io/projected/7e61a183-e84b-4e3a-8dfb-285670b4e290-kube-api-access-lw26f\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.722958 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.723114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-policies\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.723726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.723725 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e61a183-e84b-4e3a-8dfb-285670b4e290-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.723932 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.724207 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.724333 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.724799 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw4lw\" (UniqueName: \"kubernetes.io/projected/33d7a759-31d9-4a9c-a613-30db8dc0f3fe-kube-api-access-vw4lw\") pod \"downloads-7954f5f757-vrf24\" (UID: \"33d7a759-31d9-4a9c-a613-30db8dc0f3fe\") " pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.724903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-audit-policies\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.724941 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.724968 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e61a183-e84b-4e3a-8dfb-285670b4e290-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725001 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcfbdf4d-1139-4153-b4ed-783e13867f47-audit-dir\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wj72\" (UniqueName: \"kubernetes.io/projected/3f118cab-ae82-49c9-90fb-e00688b70c5b-kube-api-access-2wj72\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725087 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcfbdf4d-1139-4153-b4ed-783e13867f47-audit-dir\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725121 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725302 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725497 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcfbdf4d-1139-4153-b4ed-783e13867f47-audit-policies\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.725514 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.726371 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-serving-cert\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.726415 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-encryption-config\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.726949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.730497 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/462c99b5-fcc5-47ae-b803-9c49a676afdc-serving-cert\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.730750 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fcfbdf4d-1139-4153-b4ed-783e13867f47-etcd-client\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.731004 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.731226 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.731639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.732195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.732253 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e61a183-e84b-4e3a-8dfb-285670b4e290-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.733238 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.735894 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.756242 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.775957 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.796855 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.815772 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.835676 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.855998 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.875899 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.895857 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.916284 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.935883 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.955666 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.976272 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 14:00:57 crc kubenswrapper[4715]: I1204 14:00:57.996465 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.015728 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.035491 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.055747 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.076018 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.097180 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.115542 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.136197 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.155236 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.176050 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.195418 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.221946 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.235795 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.255530 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.275178 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.295610 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.316056 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.336280 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.356653 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.376927 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.416958 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.436605 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.456510 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.489973 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2gq\" (UniqueName: \"kubernetes.io/projected/261ac4be-400a-48b6-98de-ded0afc94441-kube-api-access-ld2gq\") pod \"controller-manager-879f6c89f-hb2z4\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.510351 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7b1d7a9-482e-4058-8a97-ae478860ee39-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.515454 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.535913 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.555875 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.575951 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.596904 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.633643 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfwt\" (UniqueName: \"kubernetes.io/projected/b1769830-312d-4f25-870a-b8aaff8b9d0d-kube-api-access-hvfwt\") pod \"machine-approver-56656f9798-6qsj9\" (UID: \"b1769830-312d-4f25-870a-b8aaff8b9d0d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.652176 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2hpb\" (UniqueName: \"kubernetes.io/projected/96c2ac59-2838-44ca-b4d9-e75092c16301-kube-api-access-h2hpb\") pod \"cluster-samples-operator-665b6dd947-fk5j5\" (UID: \"96c2ac59-2838-44ca-b4d9-e75092c16301\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.654898 4715 request.go:700] Waited for 1.015062057s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.673370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m6p5\" (UniqueName: \"kubernetes.io/projected/56596b16-b212-4e17-92ae-0050b15a5382-kube-api-access-9m6p5\") pod \"apiserver-76f77b778f-qzhn8\" (UID: \"56596b16-b212-4e17-92ae-0050b15a5382\") " pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.692846 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j45m4\" (UniqueName: \"kubernetes.io/projected/e7b1d7a9-482e-4058-8a97-ae478860ee39-kube-api-access-j45m4\") pod \"cluster-image-registry-operator-dc59b4c8b-spmzg\" (UID: \"e7b1d7a9-482e-4058-8a97-ae478860ee39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.700173 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.713463 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9h49\" (UniqueName: \"kubernetes.io/projected/c3b08994-cb44-48a9-87d1-af33d2b511b6-kube-api-access-n9h49\") pod \"console-f9d7485db-vvwbg\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.731962 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.732853 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs56r\" (UniqueName: \"kubernetes.io/projected/f68f6c4b-010e-47f0-9a1f-74facb7f8190-kube-api-access-bs56r\") pod \"authentication-operator-69f744f599-wmltk\" (UID: \"f68f6c4b-010e-47f0-9a1f-74facb7f8190\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.746067 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.749966 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gnlf\" (UniqueName: \"kubernetes.io/projected/0cf11e59-8054-46a5-8a2d-f7c6434a6a91-kube-api-access-8gnlf\") pod \"machine-api-operator-5694c8668f-c8mr7\" (UID: \"0cf11e59-8054-46a5-8a2d-f7c6434a6a91\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.771417 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7z2k\" (UniqueName: \"kubernetes.io/projected/eebcc6c0-a71c-48c5-bd62-05c09c4301d7-kube-api-access-m7z2k\") pod \"openshift-controller-manager-operator-756b6f6bc6-dkrwd\" (UID: \"eebcc6c0-a71c-48c5-bd62-05c09c4301d7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.771645 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.782738 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.797380 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.799358 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.813069 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.816410 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.836512 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.856153 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.862339 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.874682 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.876733 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.896166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.916468 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.937883 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.957123 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.969495 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.977600 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 14:00:58 crc kubenswrapper[4715]: I1204 14:00:58.996572 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.014065 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.017263 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.036704 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.056405 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.075683 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.090979 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.099474 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.109712 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hb2z4"] Dec 04 14:00:59 crc kubenswrapper[4715]: W1204 14:00:59.111971 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7b1d7a9_482e_4058_8a97_ae478860ee39.slice/crio-c2200e4a11daac34bc680a60190101f05d0d0b8abb091d5dac50ef17eaed305e WatchSource:0}: Error finding container c2200e4a11daac34bc680a60190101f05d0d0b8abb091d5dac50ef17eaed305e: Status 404 returned error can't find the container with id c2200e4a11daac34bc680a60190101f05d0d0b8abb091d5dac50ef17eaed305e Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.117487 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.137326 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.155639 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.163567 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qzhn8"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.228614 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.228899 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wmltk"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.242339 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.242700 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgvbc\" (UniqueName: \"kubernetes.io/projected/5e345625-b45e-479d-9bcb-ca495970b31b-kube-api-access-sgvbc\") pod \"console-operator-58897d9998-qtq4r\" (UID: \"5e345625-b45e-479d-9bcb-ca495970b31b\") " pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.247978 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdckm\" (UniqueName: \"kubernetes.io/projected/ece56c58-71e4-421a-8dc0-04a543033972-kube-api-access-rdckm\") pod \"route-controller-manager-6576b87f9c-drpkt\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.264387 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.265739 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.267686 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c8mr7"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.276587 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.278294 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" event={"ID":"b1769830-312d-4f25-870a-b8aaff8b9d0d","Type":"ContainerStarted","Data":"d22e66b776fe17e66c3f135b3bb4b0766e5d869a0564991a8a80323f60c39e38"} Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.278346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" event={"ID":"b1769830-312d-4f25-870a-b8aaff8b9d0d","Type":"ContainerStarted","Data":"48263811756d7e982744fffb1d6c5f86fa54c22e204aef6bfa24e7dbd853e8f1"} Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.279827 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" event={"ID":"e7b1d7a9-482e-4058-8a97-ae478860ee39","Type":"ContainerStarted","Data":"c2200e4a11daac34bc680a60190101f05d0d0b8abb091d5dac50ef17eaed305e"} Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.280362 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.280578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" event={"ID":"56596b16-b212-4e17-92ae-0050b15a5382","Type":"ContainerStarted","Data":"0a8886941d6ecf728b12789a46dd60982e8bf0863b8166130bd7395ef38ccabd"} Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.282743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" event={"ID":"f68f6c4b-010e-47f0-9a1f-74facb7f8190","Type":"ContainerStarted","Data":"c2a13829dcb6f3b8e846394c95c6a528c9c473d8fb2081d13381e14bfcfb3e7f"} Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.284421 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" event={"ID":"261ac4be-400a-48b6-98de-ded0afc94441","Type":"ContainerStarted","Data":"8635a6fe214643cd0ed13b5df8d35ce1e2e5c8b9f61b0d1e9323fcd2f4cb9420"} Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.284923 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vvwbg"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.296580 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.316956 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.336375 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.355936 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.376433 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.387905 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.395624 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.416234 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.435828 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.456711 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: W1204 14:00:59.468356 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeebcc6c0_a71c_48c5_bd62_05c09c4301d7.slice/crio-22bb4c8169219be8cb8c0ca21fdb74c339e4fd47cacc21153c2c498e21bc5fbe WatchSource:0}: Error finding container 22bb4c8169219be8cb8c0ca21fdb74c339e4fd47cacc21153c2c498e21bc5fbe: Status 404 returned error can't find the container with id 22bb4c8169219be8cb8c0ca21fdb74c339e4fd47cacc21153c2c498e21bc5fbe Dec 04 14:00:59 crc kubenswrapper[4715]: W1204 14:00:59.469641 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cf11e59_8054_46a5_8a2d_f7c6434a6a91.slice/crio-875f84333baf09c7ee0e53f215e8ac500beed0d897a67df6070f27697b098e8f WatchSource:0}: Error finding container 875f84333baf09c7ee0e53f215e8ac500beed0d897a67df6070f27697b098e8f: Status 404 returned error can't find the container with id 875f84333baf09c7ee0e53f215e8ac500beed0d897a67df6070f27697b098e8f Dec 04 14:00:59 crc kubenswrapper[4715]: W1204 14:00:59.471746 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b08994_cb44_48a9_87d1_af33d2b511b6.slice/crio-77d91947076b85cffb7ba970a4c5b6dc2156f3fe9f6f01e062f82681eebde3b5 WatchSource:0}: Error finding container 77d91947076b85cffb7ba970a4c5b6dc2156f3fe9f6f01e062f82681eebde3b5: Status 404 returned error can't find the container with id 77d91947076b85cffb7ba970a4c5b6dc2156f3fe9f6f01e062f82681eebde3b5 Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.477047 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.496009 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.517245 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.536780 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.560755 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.577959 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.596925 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.615477 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.646916 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.656477 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.674651 4715 request.go:700] Waited for 1.977833769s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.676562 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.694195 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt"] Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.697455 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.717012 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 14:00:59 crc kubenswrapper[4715]: W1204 14:00:59.734463 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podece56c58_71e4_421a_8dc0_04a543033972.slice/crio-9e8751e614b30ba3ebe75b217cb574036ee2cfb0148ab10cb3a56525045bba64 WatchSource:0}: Error finding container 9e8751e614b30ba3ebe75b217cb574036ee2cfb0148ab10cb3a56525045bba64: Status 404 returned error can't find the container with id 9e8751e614b30ba3ebe75b217cb574036ee2cfb0148ab10cb3a56525045bba64 Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.735807 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.737514 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qtq4r"] Dec 04 14:00:59 crc kubenswrapper[4715]: W1204 14:00:59.760496 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e345625_b45e_479d_9bcb_ca495970b31b.slice/crio-46317036b27aecd6abfb43f015be15eba74789fd2d6f4a9769b5ad53ed32d14e WatchSource:0}: Error finding container 46317036b27aecd6abfb43f015be15eba74789fd2d6f4a9769b5ad53ed32d14e: Status 404 returned error can't find the container with id 46317036b27aecd6abfb43f015be15eba74789fd2d6f4a9769b5ad53ed32d14e Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.775906 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z8dp\" (UniqueName: \"kubernetes.io/projected/fcfbdf4d-1139-4153-b4ed-783e13867f47-kube-api-access-8z8dp\") pod \"apiserver-7bbb656c7d-95skz\" (UID: \"fcfbdf4d-1139-4153-b4ed-783e13867f47\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.791236 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2xcf\" (UniqueName: \"kubernetes.io/projected/462c99b5-fcc5-47ae-b803-9c49a676afdc-kube-api-access-t2xcf\") pod \"openshift-config-operator-7777fb866f-mbdsr\" (UID: \"462c99b5-fcc5-47ae-b803-9c49a676afdc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.807175 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.818135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw26f\" (UniqueName: \"kubernetes.io/projected/7e61a183-e84b-4e3a-8dfb-285670b4e290-kube-api-access-lw26f\") pod \"openshift-apiserver-operator-796bbdcf4f-ft5tk\" (UID: \"7e61a183-e84b-4e3a-8dfb-285670b4e290\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.829473 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.832102 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw4lw\" (UniqueName: \"kubernetes.io/projected/33d7a759-31d9-4a9c-a613-30db8dc0f3fe-kube-api-access-vw4lw\") pod \"downloads-7954f5f757-vrf24\" (UID: \"33d7a759-31d9-4a9c-a613-30db8dc0f3fe\") " pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.837999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.857235 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wj72\" (UniqueName: \"kubernetes.io/projected/3f118cab-ae82-49c9-90fb-e00688b70c5b-kube-api-access-2wj72\") pod \"oauth-openshift-558db77b4-nrpl2\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.950922 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114a5f51-bf74-40bc-8aa0-79db31ad3c95-config-volume\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2ngx\" (UniqueName: \"kubernetes.io/projected/e2bc2471-bf61-4a2e-91de-31085de69d0e-kube-api-access-p2ngx\") pod \"dns-operator-744455d44c-blwbc\" (UID: \"e2bc2471-bf61-4a2e-91de-31085de69d0e\") " pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951376 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-certificates\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951394 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9314014-3943-4679-acd8-53eefbbb7f3f-trusted-ca\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-ca\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951445 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc83a185-feec-48f3-85b6-fccb114fede6-config\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11030318-d9af-4440-ab0f-2cfed1accd08-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951521 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lspg\" (UniqueName: \"kubernetes.io/projected/efd1087d-87f9-4427-9394-729d6d910017-kube-api-access-9lspg\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/678577cf-42b5-4f4a-84d2-9be8e1d7b753-installation-pull-secrets\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951554 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efd1087d-87f9-4427-9394-729d6d910017-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951584 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd077bbf-e883-4390-b866-e12661c88325-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951612 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsn5x\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-kube-api-access-dsn5x\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951630 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w76wr\" (UniqueName: \"kubernetes.io/projected/57fb6a21-7170-4eb7-b38d-057b9af45aca-kube-api-access-w76wr\") pod \"multus-admission-controller-857f4d67dd-m4qn9\" (UID: \"57fb6a21-7170-4eb7-b38d-057b9af45aca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951662 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz4np\" (UniqueName: \"kubernetes.io/projected/0625d627-3c69-4b5e-815a-0a3659f14fb4-kube-api-access-dz4np\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951679 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8w5r\" (UniqueName: \"kubernetes.io/projected/76eb3c9c-3638-4f45-8d06-b91d96f05c77-kube-api-access-m8w5r\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd077bbf-e883-4390-b866-e12661c88325-config\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11030318-d9af-4440-ab0f-2cfed1accd08-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951754 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/114a5f51-bf74-40bc-8aa0-79db31ad3c95-metrics-tls\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951773 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-proxy-tls\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951832 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2bc2471-bf61-4a2e-91de-31085de69d0e-metrics-tls\") pod \"dns-operator-744455d44c-blwbc\" (UID: \"e2bc2471-bf61-4a2e-91de-31085de69d0e\") " pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951862 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/678577cf-42b5-4f4a-84d2-9be8e1d7b753-ca-trust-extracted\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a695f338-e2a0-44e5-925a-ac06e19aeaac-service-ca-bundle\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951919 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-config\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.951961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z69fc\" (UniqueName: \"kubernetes.io/projected/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-kube-api-access-z69fc\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952024 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-default-certificate\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952146 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-proxy-tls\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-bound-sa-token\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952191 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg45s\" (UniqueName: \"kubernetes.io/projected/e8b7067f-d893-45b2-8227-90725570298e-kube-api-access-vg45s\") pod \"migrator-59844c95c7-7rwtm\" (UID: \"e8b7067f-d893-45b2-8227-90725570298e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efd1087d-87f9-4427-9394-729d6d910017-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l828z\" (UniqueName: \"kubernetes.io/projected/237d1dc9-8b95-42eb-8c48-5f376ef06470-kube-api-access-l828z\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd077bbf-e883-4390-b866-e12661c88325-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952317 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc83a185-feec-48f3-85b6-fccb114fede6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952335 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/237d1dc9-8b95-42eb-8c48-5f376ef06470-tmpfs\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952372 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-stats-auth\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-images\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952452 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-trusted-ca\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jclrz\" (UniqueName: \"kubernetes.io/projected/114a5f51-bf74-40bc-8aa0-79db31ad3c95-kube-api-access-jclrz\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952483 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0625d627-3c69-4b5e-815a-0a3659f14fb4-srv-cert\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952522 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/237d1dc9-8b95-42eb-8c48-5f376ef06470-apiservice-cert\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-metrics-certs\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952598 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-tls\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c6pf\" (UniqueName: \"kubernetes.io/projected/e9314014-3943-4679-acd8-53eefbbb7f3f-kube-api-access-9c6pf\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952631 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952658 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dp8l\" (UniqueName: \"kubernetes.io/projected/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-kube-api-access-6dp8l\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952679 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9314014-3943-4679-acd8-53eefbbb7f3f-metrics-tls\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed7aa2dd-2b37-4f48-80eb-2855ac8fea13-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j2zjt\" (UID: \"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0625d627-3c69-4b5e-815a-0a3659f14fb4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms96p\" (UniqueName: \"kubernetes.io/projected/ed7aa2dd-2b37-4f48-80eb-2855ac8fea13-kube-api-access-ms96p\") pod \"package-server-manager-789f6589d5-j2zjt\" (UID: \"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952834 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76eb3c9c-3638-4f45-8d06-b91d96f05c77-serving-cert\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952862 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11030318-d9af-4440-ab0f-2cfed1accd08-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952878 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc83a185-feec-48f3-85b6-fccb114fede6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952916 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-client\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952946 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfprl\" (UniqueName: \"kubernetes.io/projected/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-kube-api-access-sfprl\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.952994 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.953060 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.953130 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4d99\" (UniqueName: \"kubernetes.io/projected/a695f338-e2a0-44e5-925a-ac06e19aeaac-kube-api-access-v4d99\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.953157 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/57fb6a21-7170-4eb7-b38d-057b9af45aca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m4qn9\" (UID: \"57fb6a21-7170-4eb7-b38d-057b9af45aca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.953173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-service-ca\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.953201 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/237d1dc9-8b95-42eb-8c48-5f376ef06470-webhook-cert\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:00:59 crc kubenswrapper[4715]: I1204 14:00:59.953219 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9314014-3943-4679-acd8-53eefbbb7f3f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:00:59 crc kubenswrapper[4715]: E1204 14:00:59.954417 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.454405748 +0000 UTC m=+217.523123963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060485 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060629 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd077bbf-e883-4390-b866-e12661c88325-config\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11030318-d9af-4440-ab0f-2cfed1accd08-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060677 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/114a5f51-bf74-40bc-8aa0-79db31ad3c95-metrics-tls\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-proxy-tls\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060728 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqf6w\" (UniqueName: \"kubernetes.io/projected/288f2069-b715-4316-ac68-c075546d04c4-kube-api-access-tqf6w\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvrxw\" (UID: \"288f2069-b715-4316-ac68-c075546d04c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/895789e8-6f6c-45b7-91f2-02d092cd49c0-srv-cert\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060779 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2bc2471-bf61-4a2e-91de-31085de69d0e-metrics-tls\") pod \"dns-operator-744455d44c-blwbc\" (UID: \"e2bc2471-bf61-4a2e-91de-31085de69d0e\") " pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/678577cf-42b5-4f4a-84d2-9be8e1d7b753-ca-trust-extracted\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060820 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/288f2069-b715-4316-ac68-c075546d04c4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvrxw\" (UID: \"288f2069-b715-4316-ac68-c075546d04c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a695f338-e2a0-44e5-925a-ac06e19aeaac-service-ca-bundle\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060864 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-config\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z69fc\" (UniqueName: \"kubernetes.io/projected/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-kube-api-access-z69fc\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060908 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-secret-volume\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.060978 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-default-certificate\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061008 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-proxy-tls\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg45s\" (UniqueName: \"kubernetes.io/projected/e8b7067f-d893-45b2-8227-90725570298e-kube-api-access-vg45s\") pod \"migrator-59844c95c7-7rwtm\" (UID: \"e8b7067f-d893-45b2-8227-90725570298e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061064 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d705ca88-b088-4ff5-afaa-791ce4b792f1-signing-key\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-bound-sa-token\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061108 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/acb14845-fd9d-44d2-9aca-d74b976ca539-node-bootstrap-token\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061129 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efd1087d-87f9-4427-9394-729d6d910017-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061151 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a1ee93-8b43-4086-9174-caa4ecbe300d-config\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061173 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l828z\" (UniqueName: \"kubernetes.io/projected/237d1dc9-8b95-42eb-8c48-5f376ef06470-kube-api-access-l828z\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-plugins-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061418 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd077bbf-e883-4390-b866-e12661c88325-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc83a185-feec-48f3-85b6-fccb114fede6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061461 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/237d1dc9-8b95-42eb-8c48-5f376ef06470-tmpfs\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061501 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-stats-auth\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.061570 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.561551421 +0000 UTC m=+217.630269726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061594 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-images\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061619 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d705ca88-b088-4ff5-afaa-791ce4b792f1-signing-cabundle\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061642 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-trusted-ca\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jclrz\" (UniqueName: \"kubernetes.io/projected/114a5f51-bf74-40bc-8aa0-79db31ad3c95-kube-api-access-jclrz\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061685 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0625d627-3c69-4b5e-815a-0a3659f14fb4-srv-cert\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061704 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-metrics-certs\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061725 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/237d1dc9-8b95-42eb-8c48-5f376ef06470-apiservice-cert\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-tls\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c6pf\" (UniqueName: \"kubernetes.io/projected/e9314014-3943-4679-acd8-53eefbbb7f3f-kube-api-access-9c6pf\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061802 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061824 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dp8l\" (UniqueName: \"kubernetes.io/projected/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-kube-api-access-6dp8l\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061846 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-registration-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18a1ee93-8b43-4086-9174-caa4ecbe300d-serving-cert\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061890 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9314014-3943-4679-acd8-53eefbbb7f3f-metrics-tls\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061917 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed7aa2dd-2b37-4f48-80eb-2855ac8fea13-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j2zjt\" (UID: \"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0625d627-3c69-4b5e-815a-0a3659f14fb4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061966 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/895789e8-6f6c-45b7-91f2-02d092cd49c0-profile-collector-cert\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.061985 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-mountpoint-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062016 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms96p\" (UniqueName: \"kubernetes.io/projected/ed7aa2dd-2b37-4f48-80eb-2855ac8fea13-kube-api-access-ms96p\") pod \"package-server-manager-789f6589d5-j2zjt\" (UID: \"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062054 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76eb3c9c-3638-4f45-8d06-b91d96f05c77-serving-cert\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11030318-d9af-4440-ab0f-2cfed1accd08-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc83a185-feec-48f3-85b6-fccb114fede6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt5q9\" (UniqueName: \"kubernetes.io/projected/18a1ee93-8b43-4086-9174-caa4ecbe300d-kube-api-access-xt5q9\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-client\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfprl\" (UniqueName: \"kubernetes.io/projected/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-kube-api-access-sfprl\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062223 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062245 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-socket-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4d99\" (UniqueName: \"kubernetes.io/projected/a695f338-e2a0-44e5-925a-ac06e19aeaac-kube-api-access-v4d99\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/57fb6a21-7170-4eb7-b38d-057b9af45aca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m4qn9\" (UID: \"57fb6a21-7170-4eb7-b38d-057b9af45aca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8m49\" (UniqueName: \"kubernetes.io/projected/acb14845-fd9d-44d2-9aca-d74b976ca539-kube-api-access-z8m49\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/237d1dc9-8b95-42eb-8c48-5f376ef06470-webhook-cert\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062381 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9314014-3943-4679-acd8-53eefbbb7f3f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062403 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-service-ca\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062428 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114a5f51-bf74-40bc-8aa0-79db31ad3c95-config-volume\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062452 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2ngx\" (UniqueName: \"kubernetes.io/projected/e2bc2471-bf61-4a2e-91de-31085de69d0e-kube-api-access-p2ngx\") pod \"dns-operator-744455d44c-blwbc\" (UID: \"e2bc2471-bf61-4a2e-91de-31085de69d0e\") " pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062474 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-certificates\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062495 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9314014-3943-4679-acd8-53eefbbb7f3f-trusted-ca\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062517 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-ca\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc83a185-feec-48f3-85b6-fccb114fede6-config\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-config\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mcxk\" (UniqueName: \"kubernetes.io/projected/d705ca88-b088-4ff5-afaa-791ce4b792f1-kube-api-access-6mcxk\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062589 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11030318-d9af-4440-ab0f-2cfed1accd08-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lspg\" (UniqueName: \"kubernetes.io/projected/efd1087d-87f9-4427-9394-729d6d910017-kube-api-access-9lspg\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062632 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/678577cf-42b5-4f4a-84d2-9be8e1d7b753-installation-pull-secrets\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062646 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efd1087d-87f9-4427-9394-729d6d910017-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efd1087d-87f9-4427-9394-729d6d910017-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062679 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/acb14845-fd9d-44d2-9aca-d74b976ca539-certs\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.063169 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd077bbf-e883-4390-b866-e12661c88325-config\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.062069 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a695f338-e2a0-44e5-925a-ac06e19aeaac-service-ca-bundle\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.066752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd077bbf-e883-4390-b866-e12661c88325-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.067256 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-trusted-ca\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.067469 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.067545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/237d1dc9-8b95-42eb-8c48-5f376ef06470-tmpfs\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.068320 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-proxy-tls\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.070418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-client\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.070941 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.071839 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-certificates\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.072894 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.57287819 +0000 UTC m=+217.641596405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.072957 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/678577cf-42b5-4f4a-84d2-9be8e1d7b753-ca-trust-extracted\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.073300 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-ca\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.073577 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0625d627-3c69-4b5e-815a-0a3659f14fb4-srv-cert\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.073913 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76eb3c9c-3638-4f45-8d06-b91d96f05c77-etcd-service-ca\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.074310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9314014-3943-4679-acd8-53eefbbb7f3f-trusted-ca\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.074511 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114a5f51-bf74-40bc-8aa0-79db31ad3c95-config-volume\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.075396 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.075580 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-stats-auth\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.075919 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.077620 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2bc2471-bf61-4a2e-91de-31085de69d0e-metrics-tls\") pod \"dns-operator-744455d44c-blwbc\" (UID: \"e2bc2471-bf61-4a2e-91de-31085de69d0e\") " pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.078748 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc83a185-feec-48f3-85b6-fccb114fede6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.079993 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc83a185-feec-48f3-85b6-fccb114fede6-config\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.081639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9314014-3943-4679-acd8-53eefbbb7f3f-metrics-tls\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.081731 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd077bbf-e883-4390-b866-e12661c88325-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.081780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc292\" (UniqueName: \"kubernetes.io/projected/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-kube-api-access-mc292\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.081812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w76wr\" (UniqueName: \"kubernetes.io/projected/57fb6a21-7170-4eb7-b38d-057b9af45aca-kube-api-access-w76wr\") pod \"multus-admission-controller-857f4d67dd-m4qn9\" (UID: \"57fb6a21-7170-4eb7-b38d-057b9af45aca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.081831 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzvjt\" (UniqueName: \"kubernetes.io/projected/f58446ff-601d-408f-bdab-604fed5c8825-kube-api-access-fzvjt\") pod \"ingress-canary-2lzhm\" (UID: \"f58446ff-601d-408f-bdab-604fed5c8825\") " pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsn5x\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-kube-api-access-dsn5x\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l25z\" (UniqueName: \"kubernetes.io/projected/895789e8-6f6c-45b7-91f2-02d092cd49c0-kube-api-access-4l25z\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz4np\" (UniqueName: \"kubernetes.io/projected/0625d627-3c69-4b5e-815a-0a3659f14fb4-kube-api-access-dz4np\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082142 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b52gs\" (UniqueName: \"kubernetes.io/projected/b567c7e6-f87d-4580-b477-bc310428fb51-kube-api-access-b52gs\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082170 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8w5r\" (UniqueName: \"kubernetes.io/projected/76eb3c9c-3638-4f45-8d06-b91d96f05c77-kube-api-access-m8w5r\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-csi-data-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082217 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f58446ff-601d-408f-bdab-604fed5c8825-cert\") pod \"ingress-canary-2lzhm\" (UID: \"f58446ff-601d-408f-bdab-604fed5c8825\") " pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082239 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-config-volume\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.082630 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/678577cf-42b5-4f4a-84d2-9be8e1d7b753-installation-pull-secrets\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.083285 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.083515 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11030318-d9af-4440-ab0f-2cfed1accd08-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.083703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-images\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.085430 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/237d1dc9-8b95-42eb-8c48-5f376ef06470-apiservice-cert\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.085555 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed7aa2dd-2b37-4f48-80eb-2855ac8fea13-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j2zjt\" (UID: \"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.086148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-default-certificate\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.087287 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-proxy-tls\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.087762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/57fb6a21-7170-4eb7-b38d-057b9af45aca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m4qn9\" (UID: \"57fb6a21-7170-4eb7-b38d-057b9af45aca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.088232 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11030318-d9af-4440-ab0f-2cfed1accd08-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.088491 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.088629 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efd1087d-87f9-4427-9394-729d6d910017-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.093207 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vrf24"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.095528 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a695f338-e2a0-44e5-925a-ac06e19aeaac-metrics-certs\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.098774 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l828z\" (UniqueName: \"kubernetes.io/projected/237d1dc9-8b95-42eb-8c48-5f376ef06470-kube-api-access-l828z\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.098807 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0625d627-3c69-4b5e-815a-0a3659f14fb4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.100548 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-tls\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.101686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76eb3c9c-3638-4f45-8d06-b91d96f05c77-serving-cert\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.105663 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/114a5f51-bf74-40bc-8aa0-79db31ad3c95-metrics-tls\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.118021 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.129696 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/237d1dc9-8b95-42eb-8c48-5f376ef06470-webhook-cert\") pod \"packageserver-d55dfcdfc-qshwf\" (UID: \"237d1dc9-8b95-42eb-8c48-5f376ef06470\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.130820 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.133991 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg45s\" (UniqueName: \"kubernetes.io/projected/e8b7067f-d893-45b2-8227-90725570298e-kube-api-access-vg45s\") pod \"migrator-59844c95c7-7rwtm\" (UID: \"e8b7067f-d893-45b2-8227-90725570298e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.143442 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-bound-sa-token\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.167808 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z69fc\" (UniqueName: \"kubernetes.io/projected/70786eb3-71ca-4d98-a076-f42f1ddb7a0d-kube-api-access-z69fc\") pod \"machine-config-controller-84d6567774-fw42m\" (UID: \"70786eb3-71ca-4d98-a076-f42f1ddb7a0d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185090 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-secret-volume\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d705ca88-b088-4ff5-afaa-791ce4b792f1-signing-key\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/acb14845-fd9d-44d2-9aca-d74b976ca539-node-bootstrap-token\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185344 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a1ee93-8b43-4086-9174-caa4ecbe300d-config\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185361 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-plugins-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d705ca88-b088-4ff5-afaa-791ce4b792f1-signing-cabundle\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-registration-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185445 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18a1ee93-8b43-4086-9174-caa4ecbe300d-serving-cert\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185473 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/895789e8-6f6c-45b7-91f2-02d092cd49c0-profile-collector-cert\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-mountpoint-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185517 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt5q9\" (UniqueName: \"kubernetes.io/projected/18a1ee93-8b43-4086-9174-caa4ecbe300d-kube-api-access-xt5q9\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-socket-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185581 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8m49\" (UniqueName: \"kubernetes.io/projected/acb14845-fd9d-44d2-9aca-d74b976ca539-kube-api-access-z8m49\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mcxk\" (UniqueName: \"kubernetes.io/projected/d705ca88-b088-4ff5-afaa-791ce4b792f1-kube-api-access-6mcxk\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185640 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/acb14845-fd9d-44d2-9aca-d74b976ca539-certs\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc292\" (UniqueName: \"kubernetes.io/projected/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-kube-api-access-mc292\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185698 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzvjt\" (UniqueName: \"kubernetes.io/projected/f58446ff-601d-408f-bdab-604fed5c8825-kube-api-access-fzvjt\") pod \"ingress-canary-2lzhm\" (UID: \"f58446ff-601d-408f-bdab-604fed5c8825\") " pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185715 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l25z\" (UniqueName: \"kubernetes.io/projected/895789e8-6f6c-45b7-91f2-02d092cd49c0-kube-api-access-4l25z\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185739 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b52gs\" (UniqueName: \"kubernetes.io/projected/b567c7e6-f87d-4580-b477-bc310428fb51-kube-api-access-b52gs\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185761 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-csi-data-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185776 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f58446ff-601d-408f-bdab-604fed5c8825-cert\") pod \"ingress-canary-2lzhm\" (UID: \"f58446ff-601d-408f-bdab-604fed5c8825\") " pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185793 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-config-volume\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185811 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqf6w\" (UniqueName: \"kubernetes.io/projected/288f2069-b715-4316-ac68-c075546d04c4-kube-api-access-tqf6w\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvrxw\" (UID: \"288f2069-b715-4316-ac68-c075546d04c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/895789e8-6f6c-45b7-91f2-02d092cd49c0-srv-cert\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185857 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/288f2069-b715-4316-ac68-c075546d04c4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvrxw\" (UID: \"288f2069-b715-4316-ac68-c075546d04c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.185913 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-registration-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.186052 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.686012447 +0000 UTC m=+217.754730722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.187407 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a1ee93-8b43-4086-9174-caa4ecbe300d-config\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.189535 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-plugins-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.189651 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-csi-data-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.189970 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-socket-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.190458 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b567c7e6-f87d-4580-b477-bc310428fb51-mountpoint-dir\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.191277 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d705ca88-b088-4ff5-afaa-791ce4b792f1-signing-cabundle\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.192088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-config-volume\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.192178 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/288f2069-b715-4316-ac68-c075546d04c4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvrxw\" (UID: \"288f2069-b715-4316-ac68-c075546d04c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.192626 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-secret-volume\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.192776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/acb14845-fd9d-44d2-9aca-d74b976ca539-node-bootstrap-token\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.199022 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d705ca88-b088-4ff5-afaa-791ce4b792f1-signing-key\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.209053 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f58446ff-601d-408f-bdab-604fed5c8825-cert\") pod \"ingress-canary-2lzhm\" (UID: \"f58446ff-601d-408f-bdab-604fed5c8825\") " pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.210122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/acb14845-fd9d-44d2-9aca-d74b976ca539-certs\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.210302 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.210620 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18a1ee93-8b43-4086-9174-caa4ecbe300d-serving-cert\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.210751 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/895789e8-6f6c-45b7-91f2-02d092cd49c0-srv-cert\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.210809 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/895789e8-6f6c-45b7-91f2-02d092cd49c0-profile-collector-cert\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.223278 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc83a185-feec-48f3-85b6-fccb114fede6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xs65j\" (UID: \"dc83a185-feec-48f3-85b6-fccb114fede6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.233360 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c6pf\" (UniqueName: \"kubernetes.io/projected/e9314014-3943-4679-acd8-53eefbbb7f3f-kube-api-access-9c6pf\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.236759 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jclrz\" (UniqueName: \"kubernetes.io/projected/114a5f51-bf74-40bc-8aa0-79db31ad3c95-kube-api-access-jclrz\") pod \"dns-default-h4k8p\" (UID: \"114a5f51-bf74-40bc-8aa0-79db31ad3c95\") " pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.263516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.273421 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfprl\" (UniqueName: \"kubernetes.io/projected/48b89c41-9f9c-42ae-8ea1-f66bb82669cf-kube-api-access-sfprl\") pod \"machine-config-operator-74547568cd-z2xgb\" (UID: \"48b89c41-9f9c-42ae-8ea1-f66bb82669cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.273783 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.286684 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.287070 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.787054744 +0000 UTC m=+217.855772949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.289493 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.282000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11030318-d9af-4440-ab0f-2cfed1accd08-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h54h5\" (UID: \"11030318-d9af-4440-ab0f-2cfed1accd08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.321561 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9314014-3943-4679-acd8-53eefbbb7f3f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qgg7h\" (UID: \"e9314014-3943-4679-acd8-53eefbbb7f3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.324126 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" event={"ID":"ece56c58-71e4-421a-8dc0-04a543033972","Type":"ContainerStarted","Data":"93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.324160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" event={"ID":"ece56c58-71e4-421a-8dc0-04a543033972","Type":"ContainerStarted","Data":"9e8751e614b30ba3ebe75b217cb574036ee2cfb0148ab10cb3a56525045bba64"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.324936 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.326263 4715 generic.go:334] "Generic (PLEG): container finished" podID="56596b16-b212-4e17-92ae-0050b15a5382" containerID="38dce54bf19596912fe83ebe42b5d33854fbcc44ce41b8aed9a7bb38c2b4f0e4" exitCode=0 Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.326307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" event={"ID":"56596b16-b212-4e17-92ae-0050b15a5382","Type":"ContainerDied","Data":"38dce54bf19596912fe83ebe42b5d33854fbcc44ce41b8aed9a7bb38c2b4f0e4"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.328094 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2ngx\" (UniqueName: \"kubernetes.io/projected/e2bc2471-bf61-4a2e-91de-31085de69d0e-kube-api-access-p2ngx\") pod \"dns-operator-744455d44c-blwbc\" (UID: \"e2bc2471-bf61-4a2e-91de-31085de69d0e\") " pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.328322 4715 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-drpkt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.328368 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" podUID="ece56c58-71e4-421a-8dc0-04a543033972" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.331087 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" event={"ID":"96c2ac59-2838-44ca-b4d9-e75092c16301","Type":"ContainerStarted","Data":"34ebcf498aafef4714837861070a5e953e62249ed11aa00fcb95919ebdee7940"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.331115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" event={"ID":"96c2ac59-2838-44ca-b4d9-e75092c16301","Type":"ContainerStarted","Data":"0d5e5adc5dc25e84b874b1f4c20122bcbd4c296c62e56b1e36f703236444013c"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.331125 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" event={"ID":"96c2ac59-2838-44ca-b4d9-e75092c16301","Type":"ContainerStarted","Data":"5e6990b34caf74086dbb4d0daef1d669d27f2f2b7dbaf9a78e5fcdbece1509c4"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.337265 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dp8l\" (UniqueName: \"kubernetes.io/projected/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-kube-api-access-6dp8l\") pod \"marketplace-operator-79b997595-q9ctx\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.338160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" event={"ID":"b1769830-312d-4f25-870a-b8aaff8b9d0d","Type":"ContainerStarted","Data":"51c0ff3166e97dacec96fafa0d5e03c019594b88b0db2171345385ab75fa5155"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.349570 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" event={"ID":"eebcc6c0-a71c-48c5-bd62-05c09c4301d7","Type":"ContainerStarted","Data":"29309f35463a7ebf04c25bc2adfcd0d9e97a1a64a45753e1ccde6b846d69cafc"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.349611 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" event={"ID":"eebcc6c0-a71c-48c5-bd62-05c09c4301d7","Type":"ContainerStarted","Data":"22bb4c8169219be8cb8c0ca21fdb74c339e4fd47cacc21153c2c498e21bc5fbe"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.357278 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" event={"ID":"f68f6c4b-010e-47f0-9a1f-74facb7f8190","Type":"ContainerStarted","Data":"35f758f955b5865d2f82294985640ac7bd5d44212e60d9378c42ca10374b93ce"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.359393 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms96p\" (UniqueName: \"kubernetes.io/projected/ed7aa2dd-2b37-4f48-80eb-2855ac8fea13-kube-api-access-ms96p\") pod \"package-server-manager-789f6589d5-j2zjt\" (UID: \"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.375154 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" event={"ID":"e7b1d7a9-482e-4058-8a97-ae478860ee39","Type":"ContainerStarted","Data":"4fa6dbc3edcdc323e5cf3e5fa23d2380c7e654bf81ef679eb1d2832d9b7d51b9"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.375865 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd077bbf-e883-4390-b866-e12661c88325-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-65xfc\" (UID: \"bd077bbf-e883-4390-b866-e12661c88325\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.387831 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.388007 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.887971688 +0000 UTC m=+217.956689903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.388578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.398976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w76wr\" (UniqueName: \"kubernetes.io/projected/57fb6a21-7170-4eb7-b38d-057b9af45aca-kube-api-access-w76wr\") pod \"multus-admission-controller-857f4d67dd-m4qn9\" (UID: \"57fb6a21-7170-4eb7-b38d-057b9af45aca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.399820 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.899464561 +0000 UTC m=+217.968182806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.406647 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" event={"ID":"261ac4be-400a-48b6-98de-ded0afc94441","Type":"ContainerStarted","Data":"b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.408411 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.422865 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.423835 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" event={"ID":"fcfbdf4d-1139-4153-b4ed-783e13867f47","Type":"ContainerStarted","Data":"178912bcfef3f4948256b93e3d2eb3636f85f6b4b92cf16e667acd816cb4f12f"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.429103 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" event={"ID":"5e345625-b45e-479d-9bcb-ca495970b31b","Type":"ContainerStarted","Data":"b543d96358940fa470c8b8fe7a0c4fa08ffe8a1dc64257ca6a34701bb860566e"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.429145 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" event={"ID":"5e345625-b45e-479d-9bcb-ca495970b31b","Type":"ContainerStarted","Data":"46317036b27aecd6abfb43f015be15eba74789fd2d6f4a9769b5ad53ed32d14e"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.430019 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.435446 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.443913 4715 patch_prober.go:28] interesting pod/console-operator-58897d9998-qtq4r container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.443968 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" podUID="5e345625-b45e-479d-9bcb-ca495970b31b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.463605 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz4np\" (UniqueName: \"kubernetes.io/projected/0625d627-3c69-4b5e-815a-0a3659f14fb4-kube-api-access-dz4np\") pod \"olm-operator-6b444d44fb-s8dsq\" (UID: \"0625d627-3c69-4b5e-815a-0a3659f14fb4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.464989 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsn5x\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-kube-api-access-dsn5x\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.465437 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.473666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8w5r\" (UniqueName: \"kubernetes.io/projected/76eb3c9c-3638-4f45-8d06-b91d96f05c77-kube-api-access-m8w5r\") pod \"etcd-operator-b45778765-n6w9b\" (UID: \"76eb3c9c-3638-4f45-8d06-b91d96f05c77\") " pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.478279 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvwbg" event={"ID":"c3b08994-cb44-48a9-87d1-af33d2b511b6","Type":"ContainerStarted","Data":"6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.478324 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvwbg" event={"ID":"c3b08994-cb44-48a9-87d1-af33d2b511b6","Type":"ContainerStarted","Data":"77d91947076b85cffb7ba970a4c5b6dc2156f3fe9f6f01e062f82681eebde3b5"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.479588 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vrf24" event={"ID":"33d7a759-31d9-4a9c-a613-30db8dc0f3fe","Type":"ContainerStarted","Data":"63cfd5e2ec5053c20b50f4cf980320c56f5fd4d542df82944e5becf39568379b"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.483089 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lspg\" (UniqueName: \"kubernetes.io/projected/efd1087d-87f9-4427-9394-729d6d910017-kube-api-access-9lspg\") pod \"kube-storage-version-migrator-operator-b67b599dd-l7d6x\" (UID: \"efd1087d-87f9-4427-9394-729d6d910017\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.485285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" event={"ID":"462c99b5-fcc5-47ae-b803-9c49a676afdc","Type":"ContainerStarted","Data":"25b54d1699fb5b7323bfce4a956af3d7a8af85294f66333d9a7218e41adb4704"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.491747 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.491945 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.492012 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.991992796 +0000 UTC m=+218.060711011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.492986 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.498304 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:00.998285998 +0000 UTC m=+218.067004223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.500833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" event={"ID":"0cf11e59-8054-46a5-8a2d-f7c6434a6a91","Type":"ContainerStarted","Data":"fb3f6869d51f2764743eed5e0b923873301de7e234c9bc057f572912abbb9d12"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.500880 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" event={"ID":"0cf11e59-8054-46a5-8a2d-f7c6434a6a91","Type":"ContainerStarted","Data":"50ad52741ad42d4c6befa6b59d236f613060e654681de38d7a82e3fc009a3d17"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.500900 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" event={"ID":"0cf11e59-8054-46a5-8a2d-f7c6434a6a91","Type":"ContainerStarted","Data":"875f84333baf09c7ee0e53f215e8ac500beed0d897a67df6070f27697b098e8f"} Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.503577 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.503734 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4d99\" (UniqueName: \"kubernetes.io/projected/a695f338-e2a0-44e5-925a-ac06e19aeaac-kube-api-access-v4d99\") pod \"router-default-5444994796-tplh9\" (UID: \"a695f338-e2a0-44e5-925a-ac06e19aeaac\") " pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: W1204 14:01:00.509677 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e61a183_e84b_4e3a_8dfb_285670b4e290.slice/crio-9b871cf6d94e486cfbf8de347bc209f980f649642db63f03770e3c5095c3b55b WatchSource:0}: Error finding container 9b871cf6d94e486cfbf8de347bc209f980f649642db63f03770e3c5095c3b55b: Status 404 returned error can't find the container with id 9b871cf6d94e486cfbf8de347bc209f980f649642db63f03770e3c5095c3b55b Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.517539 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.518570 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nrpl2"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.526733 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.534669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.537642 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc292\" (UniqueName: \"kubernetes.io/projected/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-kube-api-access-mc292\") pod \"collect-profiles-29414280-b9rwn\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.545809 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.552298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.565714 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzvjt\" (UniqueName: \"kubernetes.io/projected/f58446ff-601d-408f-bdab-604fed5c8825-kube-api-access-fzvjt\") pod \"ingress-canary-2lzhm\" (UID: \"f58446ff-601d-408f-bdab-604fed5c8825\") " pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.575434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l25z\" (UniqueName: \"kubernetes.io/projected/895789e8-6f6c-45b7-91f2-02d092cd49c0-kube-api-access-4l25z\") pod \"catalog-operator-68c6474976-87zlz\" (UID: \"895789e8-6f6c-45b7-91f2-02d092cd49c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.583282 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.598374 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.599558 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.099537581 +0000 UTC m=+218.168255796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.605205 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.615451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.634973 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b52gs\" (UniqueName: \"kubernetes.io/projected/b567c7e6-f87d-4580-b477-bc310428fb51-kube-api-access-b52gs\") pod \"csi-hostpathplugin-rfh7t\" (UID: \"b567c7e6-f87d-4580-b477-bc310428fb51\") " pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.639308 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.652950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8m49\" (UniqueName: \"kubernetes.io/projected/acb14845-fd9d-44d2-9aca-d74b976ca539-kube-api-access-z8m49\") pod \"machine-config-server-7jmdt\" (UID: \"acb14845-fd9d-44d2-9aca-d74b976ca539\") " pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.669406 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.669764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mcxk\" (UniqueName: \"kubernetes.io/projected/d705ca88-b088-4ff5-afaa-791ce4b792f1-kube-api-access-6mcxk\") pod \"service-ca-9c57cc56f-rj8j9\" (UID: \"d705ca88-b088-4ff5-afaa-791ce4b792f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.673647 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2lzhm" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.677494 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt5q9\" (UniqueName: \"kubernetes.io/projected/18a1ee93-8b43-4086-9174-caa4ecbe300d-kube-api-access-xt5q9\") pod \"service-ca-operator-777779d784-kppt9\" (UID: \"18a1ee93-8b43-4086-9174-caa4ecbe300d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.714320 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.748239 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.248215468 +0000 UTC m=+218.316933683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.750909 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqf6w\" (UniqueName: \"kubernetes.io/projected/288f2069-b715-4316-ac68-c075546d04c4-kube-api-access-tqf6w\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvrxw\" (UID: \"288f2069-b715-4316-ac68-c075546d04c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.753318 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.786656 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.787080 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.793707 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.824305 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf"] Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.851320 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.852150 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.352131203 +0000 UTC m=+218.420849418 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.918315 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.936163 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.936752 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7jmdt" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.951325 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" Dec 04 14:01:00 crc kubenswrapper[4715]: I1204 14:01:00.953757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:00 crc kubenswrapper[4715]: E1204 14:01:00.959344 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.459321798 +0000 UTC m=+218.528040013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: W1204 14:01:01.024264 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod237d1dc9_8b95_42eb_8c48_5f376ef06470.slice/crio-8370b3984e556d1eb65e1684bafc7dffb8000768e31b38388144863298bcd8c7 WatchSource:0}: Error finding container 8370b3984e556d1eb65e1684bafc7dffb8000768e31b38388144863298bcd8c7: Status 404 returned error can't find the container with id 8370b3984e556d1eb65e1684bafc7dffb8000768e31b38388144863298bcd8c7 Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.055834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.056263 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.556244143 +0000 UTC m=+218.624962358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.135885 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.145143 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.162762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.163113 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.663101968 +0000 UTC m=+218.731820183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: W1204 14:01:01.246487 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc83a185_feec_48f3_85b6_fccb114fede6.slice/crio-2b5aa80386c56c326fff470feb16d65a55ac0baa6bb1aeff2c76187418b31256 WatchSource:0}: Error finding container 2b5aa80386c56c326fff470feb16d65a55ac0baa6bb1aeff2c76187418b31256: Status 404 returned error can't find the container with id 2b5aa80386c56c326fff470feb16d65a55ac0baa6bb1aeff2c76187418b31256 Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.265583 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.266314 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.766296294 +0000 UTC m=+218.835014499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.278026 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" podStartSLOduration=190.278005074 podStartE2EDuration="3m10.278005074s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:01.277944522 +0000 UTC m=+218.346662737" watchObservedRunningTime="2025-12-04 14:01:01.278005074 +0000 UTC m=+218.346723289" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.299260 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m4qn9"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.369455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.369767 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.869757438 +0000 UTC m=+218.938475643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.444930 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6qsj9" podStartSLOduration=190.444895678 podStartE2EDuration="3m10.444895678s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:01.440414206 +0000 UTC m=+218.509132411" watchObservedRunningTime="2025-12-04 14:01:01.444895678 +0000 UTC m=+218.513613883" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.470758 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.471078 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:01.971063512 +0000 UTC m=+219.039781727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.520748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" event={"ID":"e8b7067f-d893-45b2-8227-90725570298e","Type":"ContainerStarted","Data":"73f3bdde13d8ae49951dee29bc100e395d0a96b1a8cccde34b5d22d900d0b27f"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.521392 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" event={"ID":"dc83a185-feec-48f3-85b6-fccb114fede6","Type":"ContainerStarted","Data":"2b5aa80386c56c326fff470feb16d65a55ac0baa6bb1aeff2c76187418b31256"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.522028 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" event={"ID":"237d1dc9-8b95-42eb-8c48-5f376ef06470","Type":"ContainerStarted","Data":"8370b3984e556d1eb65e1684bafc7dffb8000768e31b38388144863298bcd8c7"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.522985 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vrf24" event={"ID":"33d7a759-31d9-4a9c-a613-30db8dc0f3fe","Type":"ContainerStarted","Data":"922a7ddab5807083f34b93c0952fa9619801fdbea77f75fd06800138c274e385"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.523696 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.525367 4715 generic.go:334] "Generic (PLEG): container finished" podID="fcfbdf4d-1139-4153-b4ed-783e13867f47" containerID="3374547bc581dd3e0bbfec186e00605c0fa2b96ae79ef90d9c8409fb80cfb0f7" exitCode=0 Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.525425 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" event={"ID":"fcfbdf4d-1139-4153-b4ed-783e13867f47","Type":"ContainerDied","Data":"3374547bc581dd3e0bbfec186e00605c0fa2b96ae79ef90d9c8409fb80cfb0f7"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.535135 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" event={"ID":"3f118cab-ae82-49c9-90fb-e00688b70c5b","Type":"ContainerStarted","Data":"b0fb0ad986bdb9309c4210e041607c67ba279b4a35c699341dd4386bf7b11d97"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.541943 4715 generic.go:334] "Generic (PLEG): container finished" podID="462c99b5-fcc5-47ae-b803-9c49a676afdc" containerID="ad7c275479a1b8907693fc210844a0a7dcfd3447cadef6b07e2f3aaac32d36c2" exitCode=0 Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.543144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" event={"ID":"462c99b5-fcc5-47ae-b803-9c49a676afdc","Type":"ContainerDied","Data":"ad7c275479a1b8907693fc210844a0a7dcfd3447cadef6b07e2f3aaac32d36c2"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.544197 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" event={"ID":"7e61a183-e84b-4e3a-8dfb-285670b4e290","Type":"ContainerStarted","Data":"9b871cf6d94e486cfbf8de347bc209f980f649642db63f03770e3c5095c3b55b"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.549767 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" event={"ID":"70786eb3-71ca-4d98-a076-f42f1ddb7a0d","Type":"ContainerStarted","Data":"89d18d46c747de1e3432b3c8be930835246eeb2dd61ff831dac24e38f65fb4c2"} Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.550776 4715 patch_prober.go:28] interesting pod/console-operator-58897d9998-qtq4r container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.550871 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" podUID="5e345625-b45e-479d-9bcb-ca495970b31b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.566936 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dkrwd" podStartSLOduration=190.566913267 podStartE2EDuration="3m10.566913267s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:01.565154259 +0000 UTC m=+218.633872474" watchObservedRunningTime="2025-12-04 14:01:01.566913267 +0000 UTC m=+218.635631502" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.573307 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.573759 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.073727433 +0000 UTC m=+219.142445648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.656107 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-vrf24 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.656246 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vrf24" podUID="33d7a759-31d9-4a9c-a613-30db8dc0f3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.682352 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-wmltk" podStartSLOduration=190.682309816 podStartE2EDuration="3m10.682309816s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:01.611872494 +0000 UTC m=+218.680590709" watchObservedRunningTime="2025-12-04 14:01:01.682309816 +0000 UTC m=+218.751028041" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.683732 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" podStartSLOduration=189.683724135 podStartE2EDuration="3m9.683724135s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:01.68172632 +0000 UTC m=+218.750444535" watchObservedRunningTime="2025-12-04 14:01:01.683724135 +0000 UTC m=+218.752442350" Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.683757 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.685271 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.185248676 +0000 UTC m=+219.253966891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.686612 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.697308 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-h4k8p"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.702838 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.737941 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h"] Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.793851 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.803204 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.303183075 +0000 UTC m=+219.371901290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: W1204 14:01:01.844277 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod114a5f51_bf74_40bc_8aa0_79db31ad3c95.slice/crio-bd8f912fcee0e47604594f75fdc8c434626f08e9af9b440a8a8e7057bcbc55ab WatchSource:0}: Error finding container bd8f912fcee0e47604594f75fdc8c434626f08e9af9b440a8a8e7057bcbc55ab: Status 404 returned error can't find the container with id bd8f912fcee0e47604594f75fdc8c434626f08e9af9b440a8a8e7057bcbc55ab Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.915914 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:01 crc kubenswrapper[4715]: E1204 14:01:01.916267 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.41625193 +0000 UTC m=+219.484970145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:01 crc kubenswrapper[4715]: I1204 14:01:01.967111 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.011644 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.025464 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q9ctx"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.026367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.026880 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.526866498 +0000 UTC m=+219.595584713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.047159 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-n6w9b"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.088083 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-blwbc"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.101752 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.128201 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.128533 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.628515702 +0000 UTC m=+219.697233907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.229899 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.241627 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.741609538 +0000 UTC m=+219.810327753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.245641 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fk5j5" podStartSLOduration=191.245627728 podStartE2EDuration="3m11.245627728s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.245381901 +0000 UTC m=+219.314100116" watchObservedRunningTime="2025-12-04 14:01:02.245627728 +0000 UTC m=+219.314345943" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.267338 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.271108 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rfh7t"] Dec 04 14:01:02 crc kubenswrapper[4715]: W1204 14:01:02.301642 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb567c7e6_f87d_4580_b477_bc310428fb51.slice/crio-cb18c4a556c7a325e9f9b46ff84144c650cf096fc55b18e35ad3bd2b705000e9 WatchSource:0}: Error finding container cb18c4a556c7a325e9f9b46ff84144c650cf096fc55b18e35ad3bd2b705000e9: Status 404 returned error can't find the container with id cb18c4a556c7a325e9f9b46ff84144c650cf096fc55b18e35ad3bd2b705000e9 Dec 04 14:01:02 crc kubenswrapper[4715]: W1204 14:01:02.305382 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2bc2471_bf61_4a2e_91de_31085de69d0e.slice/crio-be1191b4e9a6b682e3665f3ea1aa6895d2b0f318f1018bd99e1aefe4e86b2886 WatchSource:0}: Error finding container be1191b4e9a6b682e3665f3ea1aa6895d2b0f318f1018bd99e1aefe4e86b2886: Status 404 returned error can't find the container with id be1191b4e9a6b682e3665f3ea1aa6895d2b0f318f1018bd99e1aefe4e86b2886 Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.352410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.353192 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.853173182 +0000 UTC m=+219.921891397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.385061 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.392374 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rj8j9"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.397551 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vvwbg" podStartSLOduration=191.397530323 podStartE2EDuration="3m11.397530323s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.352619797 +0000 UTC m=+219.421338022" watchObservedRunningTime="2025-12-04 14:01:02.397530323 +0000 UTC m=+219.466248538" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.455199 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.455570 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:02.955559076 +0000 UTC m=+220.024277291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.486860 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-spmzg" podStartSLOduration=191.486837099 podStartE2EDuration="3m11.486837099s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.417670142 +0000 UTC m=+219.486388357" watchObservedRunningTime="2025-12-04 14:01:02.486837099 +0000 UTC m=+219.555555314" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.490436 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2lzhm"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.543165 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-c8mr7" podStartSLOduration=190.543149516 podStartE2EDuration="3m10.543149516s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.523411268 +0000 UTC m=+219.592129493" watchObservedRunningTime="2025-12-04 14:01:02.543149516 +0000 UTC m=+219.611867731" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.559466 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.559971 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.059951435 +0000 UTC m=+220.128669650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.575764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" event={"ID":"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c","Type":"ContainerStarted","Data":"d1549b17bcd037be46c25b264407b7d30773744d5d8977b3e56ec3a258a1051d"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.604677 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" event={"ID":"e9314014-3943-4679-acd8-53eefbbb7f3f","Type":"ContainerStarted","Data":"c9ed00ef28b0f8b7a3e2a6c2812bad70f56d0c70eaba9aa133c4a64430c3fead"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.615521 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" event={"ID":"11030318-d9af-4440-ab0f-2cfed1accd08","Type":"ContainerStarted","Data":"9234af53e02522179f619178f37d3a164085cee420dea98aad84da6fedc547db"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.620959 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.623161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" event={"ID":"70786eb3-71ca-4d98-a076-f42f1ddb7a0d","Type":"ContainerStarted","Data":"0533ab439d1d3dca6940de88f959269a4bf780e98e70ce3b3c38ab3818d39733"} Dec 04 14:01:02 crc kubenswrapper[4715]: W1204 14:01:02.631609 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod288f2069_b715_4316_ac68_c075546d04c4.slice/crio-672a602c79228ef2e864b50187f3bf2878ed961d7376cc67cb13157b173d30f4 WatchSource:0}: Error finding container 672a602c79228ef2e864b50187f3bf2878ed961d7376cc67cb13157b173d30f4: Status 404 returned error can't find the container with id 672a602c79228ef2e864b50187f3bf2878ed961d7376cc67cb13157b173d30f4 Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.633748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" event={"ID":"237d1dc9-8b95-42eb-8c48-5f376ef06470","Type":"ContainerStarted","Data":"3fadd4b61346b353ee08ded459f9221c3563bd986c0290eb08a64f8c35f30f80"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.633984 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.638893 4715 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qshwf container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.638954 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" podUID="237d1dc9-8b95-42eb-8c48-5f376ef06470" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.655662 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.655748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" event={"ID":"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13","Type":"ContainerStarted","Data":"04bb792ed7786695c54899c4cbf11656c11e390603776381a0b7fc0b4e9efea5"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.665947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.666384 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.166372228 +0000 UTC m=+220.235090443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.691492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" event={"ID":"462c99b5-fcc5-47ae-b803-9c49a676afdc","Type":"ContainerStarted","Data":"b1508f2e95652a0c898323882ffe592d7df38d031d3914ca4fa53efbd43dd856"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.692129 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.695627 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" podStartSLOduration=190.695611756 podStartE2EDuration="3m10.695611756s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.694004723 +0000 UTC m=+219.762722938" watchObservedRunningTime="2025-12-04 14:01:02.695611756 +0000 UTC m=+219.764329971" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.710879 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.713527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" event={"ID":"7e61a183-e84b-4e3a-8dfb-285670b4e290","Type":"ContainerStarted","Data":"32d29eb9a65d477542496688608247c5e9e6d9fad0411eb75746420e7c134935"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.767307 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kppt9"] Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.769910 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.771511 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.271493237 +0000 UTC m=+220.340211462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.780872 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" event={"ID":"3f118cab-ae82-49c9-90fb-e00688b70c5b","Type":"ContainerStarted","Data":"46bf7866d106c6933fa0e88e1b6be150124d0ce6e9de73aa69db3c3cba96a76c"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.781686 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.783081 4715 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nrpl2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.19:6443/healthz\": dial tcp 10.217.0.19:6443: connect: connection refused" start-of-body= Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.783113 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.19:6443/healthz\": dial tcp 10.217.0.19:6443: connect: connection refused" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.784019 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7jmdt" event={"ID":"acb14845-fd9d-44d2-9aca-d74b976ca539","Type":"ContainerStarted","Data":"55b3e2583e13cbdca9c48a0aba1598393ebf4796e4012f38402dbc797a68824c"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.784115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7jmdt" event={"ID":"acb14845-fd9d-44d2-9aca-d74b976ca539","Type":"ContainerStarted","Data":"dff66b06f499d620315e29fe58fb6aa19a3ab3d453a052d481cba7325894312b"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.802815 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" event={"ID":"e2bc2471-bf61-4a2e-91de-31085de69d0e","Type":"ContainerStarted","Data":"be1191b4e9a6b682e3665f3ea1aa6895d2b0f318f1018bd99e1aefe4e86b2886"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.806512 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" event={"ID":"48b89c41-9f9c-42ae-8ea1-f66bb82669cf","Type":"ContainerStarted","Data":"a3aa72ac007f1657453b05987d23a2b79d73e6e7735ac8502fd0eb7e7bc13a3e"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.813718 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" event={"ID":"b567c7e6-f87d-4580-b477-bc310428fb51","Type":"ContainerStarted","Data":"cb18c4a556c7a325e9f9b46ff84144c650cf096fc55b18e35ad3bd2b705000e9"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.821846 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7jmdt" podStartSLOduration=5.82182971 podStartE2EDuration="5.82182971s" podCreationTimestamp="2025-12-04 14:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.821643745 +0000 UTC m=+219.890361960" watchObservedRunningTime="2025-12-04 14:01:02.82182971 +0000 UTC m=+219.890547915" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.842464 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" event={"ID":"56596b16-b212-4e17-92ae-0050b15a5382","Type":"ContainerStarted","Data":"e06659dfdab9a8ec26f2cff39eef592d7b3019c61f00eebe77d13b0a967266f6"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.863683 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-tplh9" event={"ID":"a695f338-e2a0-44e5-925a-ac06e19aeaac","Type":"ContainerStarted","Data":"f21aacc5643029451c46342af671c3187991c2313935e4a92d4e44e186c9fdca"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.863732 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-tplh9" event={"ID":"a695f338-e2a0-44e5-925a-ac06e19aeaac","Type":"ContainerStarted","Data":"408f600c99d6d30d3b82c776f34f425393e84fa839fc5bb6e974c35fb08412f5"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.866186 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vrf24" podStartSLOduration=191.86616858 podStartE2EDuration="3m11.86616858s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.866048217 +0000 UTC m=+219.934766442" watchObservedRunningTime="2025-12-04 14:01:02.86616858 +0000 UTC m=+219.934886815" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.869466 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" event={"ID":"76eb3c9c-3638-4f45-8d06-b91d96f05c77","Type":"ContainerStarted","Data":"69dd6a0954441db7ea241bc003927c9df32b82080b54d26bf65f77ece1ec2a1a"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.871503 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.873578 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.373562622 +0000 UTC m=+220.442280937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.887267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" event={"ID":"e8b7067f-d893-45b2-8227-90725570298e","Type":"ContainerStarted","Data":"d7c6641a5ebb4d95007b4b587b3d3e946075acef90eb09d6697cc1891ed156f8"} Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.889333 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" podStartSLOduration=190.889309582 podStartE2EDuration="3m10.889309582s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.887209034 +0000 UTC m=+219.955927249" watchObservedRunningTime="2025-12-04 14:01:02.889309582 +0000 UTC m=+219.958027797" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.910228 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ft5tk" podStartSLOduration=191.910211812 podStartE2EDuration="3m11.910211812s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.909691838 +0000 UTC m=+219.978410053" watchObservedRunningTime="2025-12-04 14:01:02.910211812 +0000 UTC m=+219.978930027" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.975692 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.975805 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.475789422 +0000 UTC m=+220.544507637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.975653 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" podStartSLOduration=191.975631717 podStartE2EDuration="3m11.975631717s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.956017592 +0000 UTC m=+220.024735817" watchObservedRunningTime="2025-12-04 14:01:02.975631717 +0000 UTC m=+220.044349952" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.976083 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:02 crc kubenswrapper[4715]: E1204 14:01:02.977238 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.477223581 +0000 UTC m=+220.545941786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.982618 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" podStartSLOduration=190.982595217 podStartE2EDuration="3m10.982595217s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.975420212 +0000 UTC m=+220.044138427" watchObservedRunningTime="2025-12-04 14:01:02.982595217 +0000 UTC m=+220.051313432" Dec 04 14:01:02 crc kubenswrapper[4715]: I1204 14:01:02.993776 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" podStartSLOduration=191.993751812 podStartE2EDuration="3m11.993751812s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:02.992432636 +0000 UTC m=+220.061150851" watchObservedRunningTime="2025-12-04 14:01:02.993751812 +0000 UTC m=+220.062470037" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.027092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" event={"ID":"57fb6a21-7170-4eb7-b38d-057b9af45aca","Type":"ContainerStarted","Data":"30f94596eaabd258b97ba79a214a6ad154d749f4e2758ee3172fc5bc44708c51"} Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.056824 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" event={"ID":"bd077bbf-e883-4390-b866-e12661c88325","Type":"ContainerStarted","Data":"2350f2f9c41086363a24f2e83ebb8eec676389eecbfe50b2bfe3b4ab4c0b4f4a"} Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.071281 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-tplh9" podStartSLOduration=191.071261897 podStartE2EDuration="3m11.071261897s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:03.066130057 +0000 UTC m=+220.134848272" watchObservedRunningTime="2025-12-04 14:01:03.071261897 +0000 UTC m=+220.139980112" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.076729 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.077108 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.577087866 +0000 UTC m=+220.645806081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.101903 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" podStartSLOduration=191.101879992 podStartE2EDuration="3m11.101879992s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:03.101480171 +0000 UTC m=+220.170198386" watchObservedRunningTime="2025-12-04 14:01:03.101879992 +0000 UTC m=+220.170598217" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.116094 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" event={"ID":"0625d627-3c69-4b5e-815a-0a3659f14fb4","Type":"ContainerStarted","Data":"9cb7552f2a0e7d55b099262cac293ae6c0d50a027057c2504e0568b80350b39d"} Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.132806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-h4k8p" event={"ID":"114a5f51-bf74-40bc-8aa0-79db31ad3c95","Type":"ContainerStarted","Data":"bd8f912fcee0e47604594f75fdc8c434626f08e9af9b440a8a8e7057bcbc55ab"} Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.133440 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-vrf24 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.133490 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vrf24" podUID="33d7a759-31d9-4a9c-a613-30db8dc0f3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.179713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.181583 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.681569717 +0000 UTC m=+220.750288012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.223108 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qtq4r" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.288851 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.289228 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.789210104 +0000 UTC m=+220.857928319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.396916 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.397579 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:03.897568051 +0000 UTC m=+220.966286266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.506180 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.506495 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.006479943 +0000 UTC m=+221.075198158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.648658 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.649077 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.149064722 +0000 UTC m=+221.217782937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.754676 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.755466 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.255447095 +0000 UTC m=+221.324165310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.788011 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.794738 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:03 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:03 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:03 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.794785 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.856557 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.856955 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.356942965 +0000 UTC m=+221.425661180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:03 crc kubenswrapper[4715]: I1204 14:01:03.957197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:03 crc kubenswrapper[4715]: E1204 14:01:03.957985 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.457968451 +0000 UTC m=+221.526686666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.058837 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.059296 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.559283406 +0000 UTC m=+221.628001621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.153744 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" event={"ID":"fcfbdf4d-1139-4153-b4ed-783e13867f47","Type":"ContainerStarted","Data":"44d8be067ec1c78376f0f98f41914cc8c66778258d3a6cad1a6160aaec4f061d"} Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.160239 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.66021793 +0000 UTC m=+221.728936145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.161419 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.161920 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.162324 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.662312467 +0000 UTC m=+221.731030672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.162637 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" event={"ID":"d705ca88-b088-4ff5-afaa-791ce4b792f1","Type":"ContainerStarted","Data":"4b0b8e9f77e7650a50f1ddbc7529f5676408a0721b0b2e723aa675324019796f"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.162667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" event={"ID":"d705ca88-b088-4ff5-afaa-791ce4b792f1","Type":"ContainerStarted","Data":"555eed105faf8e4330a683974570f973491c587e88cf5fe5d80900e8f6c2dca0"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.167289 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" event={"ID":"76eb3c9c-3638-4f45-8d06-b91d96f05c77","Type":"ContainerStarted","Data":"a0f1d083fdc7f35bb259fbe47ffae83266e998bf18a9b8e2e2dc959ff5d7ae66"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.168677 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2lzhm" event={"ID":"f58446ff-601d-408f-bdab-604fed5c8825","Type":"ContainerStarted","Data":"4f2102fb97751e55524dc1d560725e5bce79a1886b45762f2a1fc21845724b6d"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.168707 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2lzhm" event={"ID":"f58446ff-601d-408f-bdab-604fed5c8825","Type":"ContainerStarted","Data":"c1fb997fcc503f419a19773dcd9faae4d16f4839b6c16dfd2440df3dee1c21c3"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.169423 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" event={"ID":"efd1087d-87f9-4427-9394-729d6d910017","Type":"ContainerStarted","Data":"87706fc3b3bdc9d80f56f663ebfd900808aef4f2f49f72386bfc44f7814216b2"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.170701 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" event={"ID":"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c","Type":"ContainerStarted","Data":"12877f99a49b339195bab7253fd1c99bf310047d18f247a6025928c2ba3e8e98"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.171807 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.172733 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" podStartSLOduration=192.172719621 podStartE2EDuration="3m12.172719621s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.172578587 +0000 UTC m=+221.241296802" watchObservedRunningTime="2025-12-04 14:01:04.172719621 +0000 UTC m=+221.241437836" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.172988 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q9ctx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.173051 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.184539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" event={"ID":"e9314014-3943-4679-acd8-53eefbbb7f3f","Type":"ContainerStarted","Data":"25548f062979cfd32085f8cebe66eb9b6192a9a96a988c9323ac3fe718f51490"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.188889 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" event={"ID":"57fb6a21-7170-4eb7-b38d-057b9af45aca","Type":"ContainerStarted","Data":"cf9f930008feae77ab399b8f8a36e7820110e48b9c9faaf1ff63c5130ec61f00"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.189772 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-n6w9b" podStartSLOduration=192.189761476 podStartE2EDuration="3m12.189761476s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.189457978 +0000 UTC m=+221.258176183" watchObservedRunningTime="2025-12-04 14:01:04.189761476 +0000 UTC m=+221.258479701" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.200425 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" event={"ID":"895789e8-6f6c-45b7-91f2-02d092cd49c0","Type":"ContainerStarted","Data":"25eff5113eb19d61142fe34aaf7973c5ea3c702092b9d1672118537ed7408a9b"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.202324 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" event={"ID":"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13","Type":"ContainerStarted","Data":"f9dafa749289b7753ab1e5a087bb5a7ccbeece2cf46878192bae34b68276f094"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.203541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" event={"ID":"0625d627-3c69-4b5e-815a-0a3659f14fb4","Type":"ContainerStarted","Data":"69f66e4a765947c33b482ce39b3459bc1505a1b7c239847b6ec4b9f46a6b094d"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.204979 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.205949 4715 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s8dsq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.206018 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" podUID="0625d627-3c69-4b5e-815a-0a3659f14fb4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.207154 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xs65j" event={"ID":"dc83a185-feec-48f3-85b6-fccb114fede6","Type":"ContainerStarted","Data":"22eebee22344d149e665a3519283ad3a16a916d83a953fe7eb9a5e90f6a4fe88"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.209892 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" event={"ID":"e8b7067f-d893-45b2-8227-90725570298e","Type":"ContainerStarted","Data":"0f7b1b9426d4c79aa367a157bfb209f44b650eb0c2a0ed61ed6bb77a9cc18cf4"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.220005 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" podStartSLOduration=192.219988241 podStartE2EDuration="3m12.219988241s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.218520461 +0000 UTC m=+221.287238676" watchObservedRunningTime="2025-12-04 14:01:04.219988241 +0000 UTC m=+221.288706456" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.220981 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-h4k8p" event={"ID":"114a5f51-bf74-40bc-8aa0-79db31ad3c95","Type":"ContainerStarted","Data":"378212aed578d87724d1682a01cd7adb344c01713db27e551d602664cbffa687"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.231331 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" event={"ID":"11030318-d9af-4440-ab0f-2cfed1accd08","Type":"ContainerStarted","Data":"7a83e6d221e8b3c08ba5dc68ed2d5cb66c17d00991099eeeaeaa63e78cf49898"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.254602 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" event={"ID":"56596b16-b212-4e17-92ae-0050b15a5382","Type":"ContainerStarted","Data":"4053282604b57d6b83c42e85e8ee2177e5ef76c4f491f69d7c8619cb5de039ff"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.258694 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rwtm" podStartSLOduration=192.258675947 podStartE2EDuration="3m12.258675947s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.257885405 +0000 UTC m=+221.326603620" watchObservedRunningTime="2025-12-04 14:01:04.258675947 +0000 UTC m=+221.327394162" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.263255 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.265441 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.765422781 +0000 UTC m=+221.834140996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.271669 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" event={"ID":"288f2069-b715-4316-ac68-c075546d04c4","Type":"ContainerStarted","Data":"7422670282fb730a42f4ac07c5b7b1df86df0da9fe5794b8907a4aabced4b8e2"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.271714 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" event={"ID":"288f2069-b715-4316-ac68-c075546d04c4","Type":"ContainerStarted","Data":"672a602c79228ef2e864b50187f3bf2878ed961d7376cc67cb13157b173d30f4"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.273116 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rj8j9" podStartSLOduration=192.27309342 podStartE2EDuration="3m12.27309342s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.239867754 +0000 UTC m=+221.308585969" watchObservedRunningTime="2025-12-04 14:01:04.27309342 +0000 UTC m=+221.341811645" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.282064 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" podStartSLOduration=192.282026474 podStartE2EDuration="3m12.282026474s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.271323972 +0000 UTC m=+221.340042187" watchObservedRunningTime="2025-12-04 14:01:04.282026474 +0000 UTC m=+221.350744679" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.298946 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-65xfc" event={"ID":"bd077bbf-e883-4390-b866-e12661c88325","Type":"ContainerStarted","Data":"cc29814e9629069c244c8c02bf38ccf9a15bae22f29d3f3701b706071c02aa2d"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.315935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" event={"ID":"18a1ee93-8b43-4086-9174-caa4ecbe300d","Type":"ContainerStarted","Data":"2c93e942bd7f6a4775bb2619b08767628446aad06dd9eb2d3b75e8ac14adba63"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.315983 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" event={"ID":"18a1ee93-8b43-4086-9174-caa4ecbe300d","Type":"ContainerStarted","Data":"691ca6628c1f89c0c110ecc9f21b6ace562fe8b8f4ed75887c1027d1befa641d"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.331991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" event={"ID":"c8d8f924-0487-4f2f-8241-4dfe2c51fb01","Type":"ContainerStarted","Data":"442b55d5629969e9be378f1a459b85bb4ea5d62a793fce3482a13594bc7f67ba"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.336354 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" event={"ID":"48b89c41-9f9c-42ae-8ea1-f66bb82669cf","Type":"ContainerStarted","Data":"be2b1f12323a87251b32b66239486a63a8993172974c93dadc5167853fbd6b7c"} Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.354535 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kppt9" podStartSLOduration=192.354516112 podStartE2EDuration="3m12.354516112s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:04.354021678 +0000 UTC m=+221.422739893" watchObservedRunningTime="2025-12-04 14:01:04.354516112 +0000 UTC m=+221.423234327" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.365122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.366481 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.866464518 +0000 UTC m=+221.935182733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.432276 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.466423 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.467812 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:04.967796043 +0000 UTC m=+222.036514258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.568105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.568376 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.068364567 +0000 UTC m=+222.137082782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.669596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.669798 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.169766724 +0000 UTC m=+222.238484939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.669913 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.670300 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.170286318 +0000 UTC m=+222.239004533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.771089 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.771781 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.271762937 +0000 UTC m=+222.340481162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.791382 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:04 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:04 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:04 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.791439 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.808349 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.808398 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.827991 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.829860 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qshwf" Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.872703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.873077 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.373057201 +0000 UTC m=+222.441775416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:04 crc kubenswrapper[4715]: I1204 14:01:04.973613 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:04 crc kubenswrapper[4715]: E1204 14:01:04.973971 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.473956745 +0000 UTC m=+222.542674960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.075062 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.075426 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.575410293 +0000 UTC m=+222.644128518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.175821 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.176016 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.675983397 +0000 UTC m=+222.744701612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.176258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.176567 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.676554903 +0000 UTC m=+222.745273118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.276762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.276925 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.776894921 +0000 UTC m=+222.845613146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.277132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.277456 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.777446506 +0000 UTC m=+222.846164721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.340605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" event={"ID":"70786eb3-71ca-4d98-a076-f42f1ddb7a0d","Type":"ContainerStarted","Data":"e2d032cbedd56ce955ad551d205c0859796e4885b2a4018d258817cd4687ebf3"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.342075 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" event={"ID":"895789e8-6f6c-45b7-91f2-02d092cd49c0","Type":"ContainerStarted","Data":"f5ae6e1d77140f52c6c7aaf6649d78ec04d4955f3b097ae0cbb7ab8c8a0afafe"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.342280 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.343402 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" event={"ID":"c8d8f924-0487-4f2f-8241-4dfe2c51fb01","Type":"ContainerStarted","Data":"ebf34ca00017234d68440996e44dcac5207890b1907b63eb25d0d6e56fd51840"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.344600 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" event={"ID":"efd1087d-87f9-4427-9394-729d6d910017","Type":"ContainerStarted","Data":"3f04aa8be264247b2621a4e211fda3b554243ae6704529d9c8ea6a7958b3274d"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.345979 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" event={"ID":"e2bc2471-bf61-4a2e-91de-31085de69d0e","Type":"ContainerStarted","Data":"220cc0d710b89e7dfb8f53e5be4f3abdd506b5e982d971acf79be923facfb660"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.346006 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" event={"ID":"e2bc2471-bf61-4a2e-91de-31085de69d0e","Type":"ContainerStarted","Data":"56875751a5f62a28e48150fdc844c4d9896db352bec480b72fde86200f265270"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.347712 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" event={"ID":"57fb6a21-7170-4eb7-b38d-057b9af45aca","Type":"ContainerStarted","Data":"346e0fdc0761692f92d6fdb9bb81f5ab388208e53e1d5263dca1ff43b59b29b4"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.349096 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" event={"ID":"48b89c41-9f9c-42ae-8ea1-f66bb82669cf","Type":"ContainerStarted","Data":"bde5e5916a5aa08771bbd6e7a3e6b182df775a64fffb683bda17769bc245a0b2"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.350061 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" event={"ID":"b567c7e6-f87d-4580-b477-bc310428fb51","Type":"ContainerStarted","Data":"64cb17bee3f278b7a6f6a6019237d6891faae4cd1b48719c990f260f7b9312e9"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.351368 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" event={"ID":"ed7aa2dd-2b37-4f48-80eb-2855ac8fea13","Type":"ContainerStarted","Data":"e7872873f64c89da6fe261d76cfbb8ce1d5180159c1271fca7b498bac1c90425"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.351470 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.352841 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-h4k8p" event={"ID":"114a5f51-bf74-40bc-8aa0-79db31ad3c95","Type":"ContainerStarted","Data":"a293e7f936e8d7973fdc819f823f0b48acf617d10e82b33f6c13dc060fa5a09c"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.352953 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.354541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" event={"ID":"e9314014-3943-4679-acd8-53eefbbb7f3f","Type":"ContainerStarted","Data":"a590512b5af7beb31ba859f52eed0d4b859699a653433a9bc817ea8fa4d8c826"} Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.356336 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q9ctx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.356382 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.357276 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.362726 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mbdsr" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.363721 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-95skz" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.365266 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8dsq" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.368540 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fw42m" podStartSLOduration=193.368494181 podStartE2EDuration="3m13.368494181s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.366699162 +0000 UTC m=+222.435417377" watchObservedRunningTime="2025-12-04 14:01:05.368494181 +0000 UTC m=+222.437212396" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.377865 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.378001 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.877977779 +0000 UTC m=+222.946695994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.378192 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.378529 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.878516074 +0000 UTC m=+222.947234289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.463368 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z2xgb" podStartSLOduration=193.463342039 podStartE2EDuration="3m13.463342039s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.393140283 +0000 UTC m=+222.461858498" watchObservedRunningTime="2025-12-04 14:01:05.463342039 +0000 UTC m=+222.532060254" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.479819 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.482138 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:05.982111361 +0000 UTC m=+223.050829576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.572940 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" podStartSLOduration=65.572902088 podStartE2EDuration="1m5.572902088s" podCreationTimestamp="2025-12-04 14:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.56159475 +0000 UTC m=+222.630312975" watchObservedRunningTime="2025-12-04 14:01:05.572902088 +0000 UTC m=+222.641620303" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.582539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.582932 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.082919302 +0000 UTC m=+223.151637517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.672146 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-blwbc" podStartSLOduration=193.672129236 podStartE2EDuration="3m13.672129236s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.671364955 +0000 UTC m=+222.740083170" watchObservedRunningTime="2025-12-04 14:01:05.672129236 +0000 UTC m=+222.740847451" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.672692 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-87zlz" podStartSLOduration=193.672688351 podStartE2EDuration="3m13.672688351s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.615301815 +0000 UTC m=+222.684020060" watchObservedRunningTime="2025-12-04 14:01:05.672688351 +0000 UTC m=+222.741406566" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.683442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.683862 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.183845296 +0000 UTC m=+223.252563511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.785293 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.785607 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.285595052 +0000 UTC m=+223.354313267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.799133 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:05 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:05 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:05 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.799188 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.802629 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-h4k8p" podStartSLOduration=8.802616407 podStartE2EDuration="8.802616407s" podCreationTimestamp="2025-12-04 14:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.749555569 +0000 UTC m=+222.818273784" watchObservedRunningTime="2025-12-04 14:01:05.802616407 +0000 UTC m=+222.871334622" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.834318 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4qn9" podStartSLOduration=193.834299971 podStartE2EDuration="3m13.834299971s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.804961181 +0000 UTC m=+222.873679396" watchObservedRunningTime="2025-12-04 14:01:05.834299971 +0000 UTC m=+222.903018176" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.835856 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qgg7h" podStartSLOduration=193.835850374 podStartE2EDuration="3m13.835850374s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.833644603 +0000 UTC m=+222.902362828" watchObservedRunningTime="2025-12-04 14:01:05.835850374 +0000 UTC m=+222.904568589" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.851849 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l7d6x" podStartSLOduration=193.85183256 podStartE2EDuration="3m13.85183256s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.849459565 +0000 UTC m=+222.918177780" watchObservedRunningTime="2025-12-04 14:01:05.85183256 +0000 UTC m=+222.920550775" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.881773 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvrxw" podStartSLOduration=193.881755596 podStartE2EDuration="3m13.881755596s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.879598157 +0000 UTC m=+222.948316372" watchObservedRunningTime="2025-12-04 14:01:05.881755596 +0000 UTC m=+222.950473821" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.886689 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.887165 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.387147023 +0000 UTC m=+223.455865238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.915486 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" podStartSLOduration=194.915473136 podStartE2EDuration="3m14.915473136s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.915338232 +0000 UTC m=+222.984056457" watchObservedRunningTime="2025-12-04 14:01:05.915473136 +0000 UTC m=+222.984191351" Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.988887 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:05 crc kubenswrapper[4715]: E1204 14:01:05.989635 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.489598649 +0000 UTC m=+223.558316934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:05 crc kubenswrapper[4715]: I1204 14:01:05.992324 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" podStartSLOduration=193.992300883 podStartE2EDuration="3m13.992300883s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:05.989487256 +0000 UTC m=+223.058205481" watchObservedRunningTime="2025-12-04 14:01:05.992300883 +0000 UTC m=+223.061019098" Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.022861 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2lzhm" podStartSLOduration=9.022844016 podStartE2EDuration="9.022844016s" podCreationTimestamp="2025-12-04 14:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:06.020945344 +0000 UTC m=+223.089663559" watchObservedRunningTime="2025-12-04 14:01:06.022844016 +0000 UTC m=+223.091562231" Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.089894 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.090300 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.59005849 +0000 UTC m=+223.658776705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.090459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.090778 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.59077089 +0000 UTC m=+223.659489105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.191734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.192095 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.692079724 +0000 UTC m=+223.760797929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.294056 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.294520 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.794498969 +0000 UTC m=+223.863217244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.365212 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" event={"ID":"b567c7e6-f87d-4580-b477-bc310428fb51","Type":"ContainerStarted","Data":"f077617ee71a82f4cd9fefec17ca55037369159ecb63b509c1ac2f3b20b562aa"} Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.366764 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q9ctx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.366817 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.395474 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.395628 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.895604808 +0000 UTC m=+223.964323023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.395845 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.396189 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.896179503 +0000 UTC m=+223.964897718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.497465 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.498883 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:06.998867885 +0000 UTC m=+224.067586100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.599622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.599966 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.099951984 +0000 UTC m=+224.168670199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.628746 4715 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.700768 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.700931 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.200910698 +0000 UTC m=+224.269628913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.701049 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.701418 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.201408192 +0000 UTC m=+224.270126407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.793315 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:06 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:06 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:06 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.793381 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.801669 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.801874 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.301845333 +0000 UTC m=+224.370563548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.801968 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.802357 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.302321036 +0000 UTC m=+224.371039321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.903240 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.903452 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.403425224 +0000 UTC m=+224.472143429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:06 crc kubenswrapper[4715]: I1204 14:01:06.903637 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:06 crc kubenswrapper[4715]: E1204 14:01:06.903945 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.403932258 +0000 UTC m=+224.472650473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.004913 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:07 crc kubenswrapper[4715]: E1204 14:01:07.005092 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.505061298 +0000 UTC m=+224.573779523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.005384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: E1204 14:01:07.005733 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.505723276 +0000 UTC m=+224.574441551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.106759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:07 crc kubenswrapper[4715]: E1204 14:01:07.106971 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.606944198 +0000 UTC m=+224.675662413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.107247 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: E1204 14:01:07.107591 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.607581015 +0000 UTC m=+224.676299230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.176953 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h54h5" podStartSLOduration=195.176935997 podStartE2EDuration="3m15.176935997s" podCreationTimestamp="2025-12-04 13:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:06.049715129 +0000 UTC m=+223.118433344" watchObservedRunningTime="2025-12-04 14:01:07.176935997 +0000 UTC m=+224.245654202" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.177485 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4tfcm"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.178567 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.181116 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.194140 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4tfcm"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.210576 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:07 crc kubenswrapper[4715]: E1204 14:01:07.211063 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.711027137 +0000 UTC m=+224.779745362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.312285 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-catalog-content\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.312386 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzhxq\" (UniqueName: \"kubernetes.io/projected/e07d8a9f-8c02-4ac6-9e82-96564833487b-kube-api-access-zzhxq\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.312413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-utilities\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.312577 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: E1204 14:01:07.313047 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 14:01:07.81301149 +0000 UTC m=+224.881729715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-79f5b" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.365935 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9kjvl"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.367127 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.369809 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.370391 4715 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T14:01:06.62877993Z","Handler":null,"Name":""} Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.373864 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" event={"ID":"b567c7e6-f87d-4580-b477-bc310428fb51","Type":"ContainerStarted","Data":"306551d893363135e272ea00bf2d4dc2cc3838972c1f624b7bc279eb64b4bef9"} Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.373904 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" event={"ID":"b567c7e6-f87d-4580-b477-bc310428fb51","Type":"ContainerStarted","Data":"b30689df71402009c7e65a7603dbf2f70f48499058b72ce1ab6deb57f8a4cc99"} Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.375952 4715 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.375994 4715 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.384451 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9kjvl"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.413724 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.413998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-catalog-content\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.414063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzhxq\" (UniqueName: \"kubernetes.io/projected/e07d8a9f-8c02-4ac6-9e82-96564833487b-kube-api-access-zzhxq\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.414090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-utilities\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.414523 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-catalog-content\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.414607 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-utilities\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.430308 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-rfh7t" podStartSLOduration=10.43028804 podStartE2EDuration="10.43028804s" podCreationTimestamp="2025-12-04 14:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:07.427018181 +0000 UTC m=+224.495736396" watchObservedRunningTime="2025-12-04 14:01:07.43028804 +0000 UTC m=+224.499006265" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.444998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzhxq\" (UniqueName: \"kubernetes.io/projected/e07d8a9f-8c02-4ac6-9e82-96564833487b-kube-api-access-zzhxq\") pod \"community-operators-4tfcm\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.457366 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.492450 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.514922 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-utilities\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.515016 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-catalog-content\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.515080 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.515134 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z86nf\" (UniqueName: \"kubernetes.io/projected/314da00a-3748-4149-b36c-8ced9d9f3018-kube-api-access-z86nf\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.519828 4715 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.519866 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.541128 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-79f5b\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.567564 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwvqj"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.568952 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.586809 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwvqj"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.617444 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-utilities\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.617926 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-catalog-content\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.618002 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z86nf\" (UniqueName: \"kubernetes.io/projected/314da00a-3748-4149-b36c-8ced9d9f3018-kube-api-access-z86nf\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.618542 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-utilities\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.618809 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-catalog-content\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.638505 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z86nf\" (UniqueName: \"kubernetes.io/projected/314da00a-3748-4149-b36c-8ced9d9f3018-kube-api-access-z86nf\") pod \"certified-operators-9kjvl\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.646210 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.681936 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.719137 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhlrl\" (UniqueName: \"kubernetes.io/projected/86f90c78-55bf-4118-b096-e9109678f1ff-kube-api-access-dhlrl\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.719208 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-catalog-content\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.719263 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-utilities\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.732159 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4tfcm"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.766510 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m2rtt"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.768455 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.783685 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m2rtt"] Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.816548 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:07 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:07 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:07 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.816605 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.820901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhlrl\" (UniqueName: \"kubernetes.io/projected/86f90c78-55bf-4118-b096-e9109678f1ff-kube-api-access-dhlrl\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.820958 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-catalog-content\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.821015 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-utilities\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.822179 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-catalog-content\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.823311 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-utilities\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.839788 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhlrl\" (UniqueName: \"kubernetes.io/projected/86f90c78-55bf-4118-b096-e9109678f1ff-kube-api-access-dhlrl\") pod \"community-operators-nwvqj\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.903410 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.922790 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-utilities\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.922867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-catalog-content\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.922892 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x5hj\" (UniqueName: \"kubernetes.io/projected/0b7a0295-de02-46f9-81f0-2dee1479af29-kube-api-access-8x5hj\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:07 crc kubenswrapper[4715]: I1204 14:01:07.937694 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-79f5b"] Dec 04 14:01:07 crc kubenswrapper[4715]: W1204 14:01:07.974524 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod678577cf_42b5_4f4a_84d2_9be8e1d7b753.slice/crio-d4a4f725cbfa51133d6af3f4682fb48638e98ca9ae915af259a2ee93cd7cc8d1 WatchSource:0}: Error finding container d4a4f725cbfa51133d6af3f4682fb48638e98ca9ae915af259a2ee93cd7cc8d1: Status 404 returned error can't find the container with id d4a4f725cbfa51133d6af3f4682fb48638e98ca9ae915af259a2ee93cd7cc8d1 Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.023865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-catalog-content\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.023907 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x5hj\" (UniqueName: \"kubernetes.io/projected/0b7a0295-de02-46f9-81f0-2dee1479af29-kube-api-access-8x5hj\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.024049 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-utilities\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.024370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-catalog-content\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.024402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-utilities\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.044700 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x5hj\" (UniqueName: \"kubernetes.io/projected/0b7a0295-de02-46f9-81f0-2dee1479af29-kube-api-access-8x5hj\") pod \"certified-operators-m2rtt\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.100759 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwvqj"] Dec 04 14:01:08 crc kubenswrapper[4715]: W1204 14:01:08.110865 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86f90c78_55bf_4118_b096_e9109678f1ff.slice/crio-97a6d27b2cc37a54c4b093fbda0ded5c7382c31b4a1ea798ce6159fa69b3d7d4 WatchSource:0}: Error finding container 97a6d27b2cc37a54c4b093fbda0ded5c7382c31b4a1ea798ce6159fa69b3d7d4: Status 404 returned error can't find the container with id 97a6d27b2cc37a54c4b093fbda0ded5c7382c31b4a1ea798ce6159fa69b3d7d4 Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.120114 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.182698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9kjvl"] Dec 04 14:01:08 crc kubenswrapper[4715]: W1204 14:01:08.192355 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod314da00a_3748_4149_b36c_8ced9d9f3018.slice/crio-afb3da6ecb02f9164f8eaab66522d931738cf6bd6dfc68b30a33ce794bd6fe2b WatchSource:0}: Error finding container afb3da6ecb02f9164f8eaab66522d931738cf6bd6dfc68b30a33ce794bd6fe2b: Status 404 returned error can't find the container with id afb3da6ecb02f9164f8eaab66522d931738cf6bd6dfc68b30a33ce794bd6fe2b Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.328138 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m2rtt"] Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.336909 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.343947 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.348511 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.360870 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.361293 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.381479 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2rtt" event={"ID":"0b7a0295-de02-46f9-81f0-2dee1479af29","Type":"ContainerStarted","Data":"79db332a5c33462b2ebcd9485ff427d39af6c90570b12f591620550a93116176"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.383459 4715 generic.go:334] "Generic (PLEG): container finished" podID="86f90c78-55bf-4118-b096-e9109678f1ff" containerID="37fc4be1d4d228abfc0721b4f0f8ce3133caf583c354f25c71be698f81b6a6d3" exitCode=0 Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.383531 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwvqj" event={"ID":"86f90c78-55bf-4118-b096-e9109678f1ff","Type":"ContainerDied","Data":"37fc4be1d4d228abfc0721b4f0f8ce3133caf583c354f25c71be698f81b6a6d3"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.383559 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwvqj" event={"ID":"86f90c78-55bf-4118-b096-e9109678f1ff","Type":"ContainerStarted","Data":"97a6d27b2cc37a54c4b093fbda0ded5c7382c31b4a1ea798ce6159fa69b3d7d4"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.390172 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" event={"ID":"678577cf-42b5-4f4a-84d2-9be8e1d7b753","Type":"ContainerStarted","Data":"49b6854642573b40395b960ad4f35c676bcb6ed7ed8dfe8ddfde8e1f1027012d"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.392450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" event={"ID":"678577cf-42b5-4f4a-84d2-9be8e1d7b753","Type":"ContainerStarted","Data":"d4a4f725cbfa51133d6af3f4682fb48638e98ca9ae915af259a2ee93cd7cc8d1"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.392587 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.392668 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kjvl" event={"ID":"314da00a-3748-4149-b36c-8ced9d9f3018","Type":"ContainerStarted","Data":"afb3da6ecb02f9164f8eaab66522d931738cf6bd6dfc68b30a33ce794bd6fe2b"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.395244 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.399346 4715 generic.go:334] "Generic (PLEG): container finished" podID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerID="282d3af7215c4d3deccb315e2eae6b1574dced3b9163a508157591dad83a25b9" exitCode=0 Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.399537 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerDied","Data":"282d3af7215c4d3deccb315e2eae6b1574dced3b9163a508157591dad83a25b9"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.399614 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerStarted","Data":"fedec658ea178aa119927685b870d722f8b576fec17859a23dfde0bd5a8142c9"} Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.431687 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.431800 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.475328 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" podStartSLOduration=197.475313626 podStartE2EDuration="3m17.475313626s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:08.47254519 +0000 UTC m=+225.541263405" watchObservedRunningTime="2025-12-04 14:01:08.475313626 +0000 UTC m=+225.544031841" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.533322 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.533403 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.533805 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.552856 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.674251 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.764779 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.765140 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.765190 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.765898 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.766005 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042" gracePeriod=600 Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.791078 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:08 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:08 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:08 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.791148 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.800072 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.800391 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.807516 4715 patch_prober.go:28] interesting pod/console-f9d7485db-vvwbg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.807631 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vvwbg" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.872551 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.970702 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.970753 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:01:08 crc kubenswrapper[4715]: I1204 14:01:08.977153 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.187141 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.367751 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ssltf"] Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.370640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.373276 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.374246 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ssltf"] Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.417601 4715 generic.go:334] "Generic (PLEG): container finished" podID="314da00a-3748-4149-b36c-8ced9d9f3018" containerID="11a29881eecd867badb076befe7c43ed8449ce9ce8a5eedc129ce5c95e3c8de3" exitCode=0 Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.417672 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kjvl" event={"ID":"314da00a-3748-4149-b36c-8ced9d9f3018","Type":"ContainerDied","Data":"11a29881eecd867badb076befe7c43ed8449ce9ce8a5eedc129ce5c95e3c8de3"} Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.421970 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042" exitCode=0 Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.422082 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042"} Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.423666 4715 generic.go:334] "Generic (PLEG): container finished" podID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerID="d2d984631c7977b509167cde50585897c8013555a9027046432308b7a8e4f8b6" exitCode=0 Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.423729 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2rtt" event={"ID":"0b7a0295-de02-46f9-81f0-2dee1479af29","Type":"ContainerDied","Data":"d2d984631c7977b509167cde50585897c8013555a9027046432308b7a8e4f8b6"} Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.427321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc","Type":"ContainerStarted","Data":"65a67df9820f42b2646c0cb84869f4915e3954b6f59c5b1f0ac6b09292aba57b"} Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.427349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc","Type":"ContainerStarted","Data":"03b3196fb97ed853e315dcccf08d45b2408c2ac1d313ec5c6b25e3fabe18131d"} Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.436277 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qzhn8" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.447958 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-catalog-content\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.448010 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhbft\" (UniqueName: \"kubernetes.io/projected/605514ff-63cc-4e34-a76e-44dd58be1c3e-kube-api-access-hhbft\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.448122 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-utilities\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.473827 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.473801012 podStartE2EDuration="1.473801012s" podCreationTimestamp="2025-12-04 14:01:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:09.470491791 +0000 UTC m=+226.539210036" watchObservedRunningTime="2025-12-04 14:01:09.473801012 +0000 UTC m=+226.542519247" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.549851 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhbft\" (UniqueName: \"kubernetes.io/projected/605514ff-63cc-4e34-a76e-44dd58be1c3e-kube-api-access-hhbft\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.550066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-utilities\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.550927 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-catalog-content\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.558394 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-catalog-content\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.558810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-utilities\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.602591 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhbft\" (UniqueName: \"kubernetes.io/projected/605514ff-63cc-4e34-a76e-44dd58be1c3e-kube-api-access-hhbft\") pod \"redhat-marketplace-ssltf\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.703857 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.774057 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-46t2d"] Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.775543 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.781826 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46t2d"] Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.791538 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:09 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:09 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:09 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.791587 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.847791 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-vrf24 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.847825 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-vrf24 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.847844 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vrf24" podUID="33d7a759-31d9-4a9c-a613-30db8dc0f3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.847878 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vrf24" podUID="33d7a759-31d9-4a9c-a613-30db8dc0f3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.857463 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-catalog-content\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.857532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxmnd\" (UniqueName: \"kubernetes.io/projected/53cf9350-c1c3-4557-8c23-2084355e0646-kube-api-access-lxmnd\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.857579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-utilities\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.960657 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-utilities\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.962006 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-catalog-content\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.961995 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-utilities\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.961197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-catalog-content\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.962196 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxmnd\" (UniqueName: \"kubernetes.io/projected/53cf9350-c1c3-4557-8c23-2084355e0646-kube-api-access-lxmnd\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:09 crc kubenswrapper[4715]: I1204 14:01:09.985155 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxmnd\" (UniqueName: \"kubernetes.io/projected/53cf9350-c1c3-4557-8c23-2084355e0646-kube-api-access-lxmnd\") pod \"redhat-marketplace-46t2d\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.117441 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ssltf"] Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.121625 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:01:10 crc kubenswrapper[4715]: W1204 14:01:10.140005 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod605514ff_63cc_4e34_a76e_44dd58be1c3e.slice/crio-39cda0e41e6f49ff01a4d9e97457191b8533286596b866ec5e1099457601d755 WatchSource:0}: Error finding container 39cda0e41e6f49ff01a4d9e97457191b8533286596b866ec5e1099457601d755: Status 404 returned error can't find the container with id 39cda0e41e6f49ff01a4d9e97457191b8533286596b866ec5e1099457601d755 Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.375146 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vkbdv"] Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.376445 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.382397 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.409132 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkbdv"] Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.417881 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46t2d"] Dec 04 14:01:10 crc kubenswrapper[4715]: W1204 14:01:10.465186 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53cf9350_c1c3_4557_8c23_2084355e0646.slice/crio-c7b64722f3dbb7cfd6dc4f774071f34c0a3c0c02cd57927dc4404944684bb320 WatchSource:0}: Error finding container c7b64722f3dbb7cfd6dc4f774071f34c0a3c0c02cd57927dc4404944684bb320: Status 404 returned error can't find the container with id c7b64722f3dbb7cfd6dc4f774071f34c0a3c0c02cd57927dc4404944684bb320 Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.469342 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"a679b434089a9d4f4ce6522120aa8528b8358549b266643a6b526d3fdfd66568"} Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.476770 4715 generic.go:334] "Generic (PLEG): container finished" podID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerID="c496fdaa4ee89b6db33e36b8dc3021a856e12f1bf6fb77232aba218cada6690d" exitCode=0 Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.476833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ssltf" event={"ID":"605514ff-63cc-4e34-a76e-44dd58be1c3e","Type":"ContainerDied","Data":"c496fdaa4ee89b6db33e36b8dc3021a856e12f1bf6fb77232aba218cada6690d"} Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.476857 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ssltf" event={"ID":"605514ff-63cc-4e34-a76e-44dd58be1c3e","Type":"ContainerStarted","Data":"39cda0e41e6f49ff01a4d9e97457191b8533286596b866ec5e1099457601d755"} Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.478417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78qvb\" (UniqueName: \"kubernetes.io/projected/b1fb6e69-52de-41b9-acae-df242cab542e-kube-api-access-78qvb\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.478508 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-catalog-content\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.478538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-utilities\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.482779 4715 generic.go:334] "Generic (PLEG): container finished" podID="c8d8f924-0487-4f2f-8241-4dfe2c51fb01" containerID="ebf34ca00017234d68440996e44dcac5207890b1907b63eb25d0d6e56fd51840" exitCode=0 Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.482833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" event={"ID":"c8d8f924-0487-4f2f-8241-4dfe2c51fb01","Type":"ContainerDied","Data":"ebf34ca00017234d68440996e44dcac5207890b1907b63eb25d0d6e56fd51840"} Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.487197 4715 generic.go:334] "Generic (PLEG): container finished" podID="3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc" containerID="65a67df9820f42b2646c0cb84869f4915e3954b6f59c5b1f0ac6b09292aba57b" exitCode=0 Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.490412 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc","Type":"ContainerDied","Data":"65a67df9820f42b2646c0cb84869f4915e3954b6f59c5b1f0ac6b09292aba57b"} Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.558647 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.580277 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78qvb\" (UniqueName: \"kubernetes.io/projected/b1fb6e69-52de-41b9-acae-df242cab542e-kube-api-access-78qvb\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.580852 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-catalog-content\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.580978 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-utilities\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.583431 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-catalog-content\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.584346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-utilities\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.612664 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78qvb\" (UniqueName: \"kubernetes.io/projected/b1fb6e69-52de-41b9-acae-df242cab542e-kube-api-access-78qvb\") pod \"redhat-operators-vkbdv\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.766343 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.774650 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p8xqs"] Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.783294 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.790877 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p8xqs"] Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.792225 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.801675 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:10 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:10 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:10 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.801797 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.909361 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-utilities\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.909405 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-catalog-content\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.909496 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fntk\" (UniqueName: \"kubernetes.io/projected/8414683b-d211-4ea5-8927-f0252a960c72-kube-api-access-6fntk\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:10 crc kubenswrapper[4715]: I1204 14:01:10.996995 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkbdv"] Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.010960 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-catalog-content\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.011000 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-utilities\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.011065 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fntk\" (UniqueName: \"kubernetes.io/projected/8414683b-d211-4ea5-8927-f0252a960c72-kube-api-access-6fntk\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.011873 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-catalog-content\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.011887 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-utilities\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.032014 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fntk\" (UniqueName: \"kubernetes.io/projected/8414683b-d211-4ea5-8927-f0252a960c72-kube-api-access-6fntk\") pod \"redhat-operators-p8xqs\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.134066 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.503581 4715 generic.go:334] "Generic (PLEG): container finished" podID="53cf9350-c1c3-4557-8c23-2084355e0646" containerID="ca17907fe2b8cc4e1c47f5217874b0a7a0dde71fb7a9434f6322f764719f52db" exitCode=0 Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.503973 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46t2d" event={"ID":"53cf9350-c1c3-4557-8c23-2084355e0646","Type":"ContainerDied","Data":"ca17907fe2b8cc4e1c47f5217874b0a7a0dde71fb7a9434f6322f764719f52db"} Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.504007 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46t2d" event={"ID":"53cf9350-c1c3-4557-8c23-2084355e0646","Type":"ContainerStarted","Data":"c7b64722f3dbb7cfd6dc4f774071f34c0a3c0c02cd57927dc4404944684bb320"} Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.506328 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p8xqs"] Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.536692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerStarted","Data":"b97a535085ff9e6399af9ef198d1c8ff3fe49acdd103cb2dd661093e9481d92b"} Dec 04 14:01:11 crc kubenswrapper[4715]: W1204 14:01:11.590190 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8414683b_d211_4ea5_8927_f0252a960c72.slice/crio-257f93f61543427eaa8dfca9335020a92813a0d074123e1ecaa603e33f9f9780 WatchSource:0}: Error finding container 257f93f61543427eaa8dfca9335020a92813a0d074123e1ecaa603e33f9f9780: Status 404 returned error can't find the container with id 257f93f61543427eaa8dfca9335020a92813a0d074123e1ecaa603e33f9f9780 Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.667978 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.671723 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.675678 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.680437 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.704691 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.791111 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:11 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:11 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:11 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.791154 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.840276 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96756a69-15d9-40a7-a8d1-3600cc193e46-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.840505 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96756a69-15d9-40a7-a8d1-3600cc193e46-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.901401 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.934535 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.947442 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96756a69-15d9-40a7-a8d1-3600cc193e46-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.947578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96756a69-15d9-40a7-a8d1-3600cc193e46-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.947794 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96756a69-15d9-40a7-a8d1-3600cc193e46-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:11 crc kubenswrapper[4715]: I1204 14:01:11.982400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96756a69-15d9-40a7-a8d1-3600cc193e46-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.023619 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.049106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kubelet-dir\") pod \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.049176 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc292\" (UniqueName: \"kubernetes.io/projected/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-kube-api-access-mc292\") pod \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.049241 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-secret-volume\") pod \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.049265 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-config-volume\") pod \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\" (UID: \"c8d8f924-0487-4f2f-8241-4dfe2c51fb01\") " Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.049303 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kube-api-access\") pod \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\" (UID: \"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc\") " Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.049946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc" (UID: "3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.050721 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-config-volume" (OuterVolumeSpecName: "config-volume") pod "c8d8f924-0487-4f2f-8241-4dfe2c51fb01" (UID: "c8d8f924-0487-4f2f-8241-4dfe2c51fb01"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.054477 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc" (UID: "3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.057226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c8d8f924-0487-4f2f-8241-4dfe2c51fb01" (UID: "c8d8f924-0487-4f2f-8241-4dfe2c51fb01"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.066652 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-kube-api-access-mc292" (OuterVolumeSpecName: "kube-api-access-mc292") pod "c8d8f924-0487-4f2f-8241-4dfe2c51fb01" (UID: "c8d8f924-0487-4f2f-8241-4dfe2c51fb01"). InnerVolumeSpecName "kube-api-access-mc292". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.150858 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.150896 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.150906 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.150916 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc292\" (UniqueName: \"kubernetes.io/projected/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-kube-api-access-mc292\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.150927 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8d8f924-0487-4f2f-8241-4dfe2c51fb01-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.383528 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 14:01:12 crc kubenswrapper[4715]: W1204 14:01:12.404469 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod96756a69_15d9_40a7_a8d1_3600cc193e46.slice/crio-47031fb336849f815719093f6ecf72a5408e4205e9aaf86e1d0de137c8d7b702 WatchSource:0}: Error finding container 47031fb336849f815719093f6ecf72a5408e4205e9aaf86e1d0de137c8d7b702: Status 404 returned error can't find the container with id 47031fb336849f815719093f6ecf72a5408e4205e9aaf86e1d0de137c8d7b702 Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.563253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerStarted","Data":"257f93f61543427eaa8dfca9335020a92813a0d074123e1ecaa603e33f9f9780"} Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.565734 4715 generic.go:334] "Generic (PLEG): container finished" podID="b1fb6e69-52de-41b9-acae-df242cab542e" containerID="590d617354d567f122b71bdbb6e864319c22f447eaee003607db67a6c9286ec2" exitCode=0 Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.566327 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerDied","Data":"590d617354d567f122b71bdbb6e864319c22f447eaee003607db67a6c9286ec2"} Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.568114 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"96756a69-15d9-40a7-a8d1-3600cc193e46","Type":"ContainerStarted","Data":"47031fb336849f815719093f6ecf72a5408e4205e9aaf86e1d0de137c8d7b702"} Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.586821 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.586823 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn" event={"ID":"c8d8f924-0487-4f2f-8241-4dfe2c51fb01","Type":"ContainerDied","Data":"442b55d5629969e9be378f1a459b85bb4ea5d62a793fce3482a13594bc7f67ba"} Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.587481 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="442b55d5629969e9be378f1a459b85bb4ea5d62a793fce3482a13594bc7f67ba" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.591456 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc","Type":"ContainerDied","Data":"03b3196fb97ed853e315dcccf08d45b2408c2ac1d313ec5c6b25e3fabe18131d"} Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.591509 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.591529 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03b3196fb97ed853e315dcccf08d45b2408c2ac1d313ec5c6b25e3fabe18131d" Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.790717 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:12 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:12 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:12 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:12 crc kubenswrapper[4715]: I1204 14:01:12.790784 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:13 crc kubenswrapper[4715]: I1204 14:01:13.617327 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerStarted","Data":"2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474"} Dec 04 14:01:13 crc kubenswrapper[4715]: I1204 14:01:13.792554 4715 patch_prober.go:28] interesting pod/router-default-5444994796-tplh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 14:01:13 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 04 14:01:13 crc kubenswrapper[4715]: [+]process-running ok Dec 04 14:01:13 crc kubenswrapper[4715]: healthz check failed Dec 04 14:01:13 crc kubenswrapper[4715]: I1204 14:01:13.792801 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tplh9" podUID="a695f338-e2a0-44e5-925a-ac06e19aeaac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 14:01:14 crc kubenswrapper[4715]: I1204 14:01:14.654170 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"96756a69-15d9-40a7-a8d1-3600cc193e46","Type":"ContainerStarted","Data":"8f4066149660c6739adf19e7d755950ed4848afc09c63b1a326c0b1abdcdaafa"} Dec 04 14:01:14 crc kubenswrapper[4715]: I1204 14:01:14.662508 4715 generic.go:334] "Generic (PLEG): container finished" podID="8414683b-d211-4ea5-8927-f0252a960c72" containerID="2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474" exitCode=0 Dec 04 14:01:14 crc kubenswrapper[4715]: I1204 14:01:14.662565 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerDied","Data":"2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474"} Dec 04 14:01:14 crc kubenswrapper[4715]: I1204 14:01:14.793828 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:14 crc kubenswrapper[4715]: I1204 14:01:14.796550 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-tplh9" Dec 04 14:01:15 crc kubenswrapper[4715]: I1204 14:01:15.469167 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-h4k8p" Dec 04 14:01:15 crc kubenswrapper[4715]: I1204 14:01:15.698475 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.698456662 podStartE2EDuration="4.698456662s" podCreationTimestamp="2025-12-04 14:01:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:15.68261008 +0000 UTC m=+232.751328295" watchObservedRunningTime="2025-12-04 14:01:15.698456662 +0000 UTC m=+232.767174877" Dec 04 14:01:16 crc kubenswrapper[4715]: I1204 14:01:16.675272 4715 generic.go:334] "Generic (PLEG): container finished" podID="96756a69-15d9-40a7-a8d1-3600cc193e46" containerID="8f4066149660c6739adf19e7d755950ed4848afc09c63b1a326c0b1abdcdaafa" exitCode=0 Dec 04 14:01:16 crc kubenswrapper[4715]: I1204 14:01:16.675353 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"96756a69-15d9-40a7-a8d1-3600cc193e46","Type":"ContainerDied","Data":"8f4066149660c6739adf19e7d755950ed4848afc09c63b1a326c0b1abdcdaafa"} Dec 04 14:01:18 crc kubenswrapper[4715]: I1204 14:01:18.800454 4715 patch_prober.go:28] interesting pod/console-f9d7485db-vvwbg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 04 14:01:18 crc kubenswrapper[4715]: I1204 14:01:18.800806 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vvwbg" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 04 14:01:19 crc kubenswrapper[4715]: I1204 14:01:19.856364 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vrf24" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.065762 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.151197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96756a69-15d9-40a7-a8d1-3600cc193e46-kubelet-dir\") pod \"96756a69-15d9-40a7-a8d1-3600cc193e46\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.151284 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96756a69-15d9-40a7-a8d1-3600cc193e46-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "96756a69-15d9-40a7-a8d1-3600cc193e46" (UID: "96756a69-15d9-40a7-a8d1-3600cc193e46"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.151322 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96756a69-15d9-40a7-a8d1-3600cc193e46-kube-api-access\") pod \"96756a69-15d9-40a7-a8d1-3600cc193e46\" (UID: \"96756a69-15d9-40a7-a8d1-3600cc193e46\") " Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.151622 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96756a69-15d9-40a7-a8d1-3600cc193e46-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.162758 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96756a69-15d9-40a7-a8d1-3600cc193e46-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "96756a69-15d9-40a7-a8d1-3600cc193e46" (UID: "96756a69-15d9-40a7-a8d1-3600cc193e46"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.253428 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96756a69-15d9-40a7-a8d1-3600cc193e46-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.728610 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"96756a69-15d9-40a7-a8d1-3600cc193e46","Type":"ContainerDied","Data":"47031fb336849f815719093f6ecf72a5408e4205e9aaf86e1d0de137c8d7b702"} Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.728651 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47031fb336849f815719093f6ecf72a5408e4205e9aaf86e1d0de137c8d7b702" Dec 04 14:01:25 crc kubenswrapper[4715]: I1204 14:01:25.728704 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 14:01:27 crc kubenswrapper[4715]: I1204 14:01:27.652073 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:01:28 crc kubenswrapper[4715]: I1204 14:01:28.805199 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:01:28 crc kubenswrapper[4715]: I1204 14:01:28.808302 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:01:40 crc kubenswrapper[4715]: I1204 14:01:40.614065 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j2zjt" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.054129 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 14:01:45 crc kubenswrapper[4715]: E1204 14:01:45.054867 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d8f924-0487-4f2f-8241-4dfe2c51fb01" containerName="collect-profiles" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.054879 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d8f924-0487-4f2f-8241-4dfe2c51fb01" containerName="collect-profiles" Dec 04 14:01:45 crc kubenswrapper[4715]: E1204 14:01:45.054899 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96756a69-15d9-40a7-a8d1-3600cc193e46" containerName="pruner" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.054905 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="96756a69-15d9-40a7-a8d1-3600cc193e46" containerName="pruner" Dec 04 14:01:45 crc kubenswrapper[4715]: E1204 14:01:45.054913 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc" containerName="pruner" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.054919 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc" containerName="pruner" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.055007 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="96756a69-15d9-40a7-a8d1-3600cc193e46" containerName="pruner" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.055019 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fdbb6d7-ea54-4a9b-91d6-9be42c6bd1bc" containerName="pruner" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.055029 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d8f924-0487-4f2f-8241-4dfe2c51fb01" containerName="collect-profiles" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.055451 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.055542 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.058289 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.058633 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.060296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e478be46-943d-4933-8cbb-b6f577e7f164-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.060358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e478be46-943d-4933-8cbb-b6f577e7f164-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.161664 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e478be46-943d-4933-8cbb-b6f577e7f164-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.161751 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e478be46-943d-4933-8cbb-b6f577e7f164-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.161799 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e478be46-943d-4933-8cbb-b6f577e7f164-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.178937 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e478be46-943d-4933-8cbb-b6f577e7f164-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:45 crc kubenswrapper[4715]: I1204 14:01:45.372975 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:47 crc kubenswrapper[4715]: E1204 14:01:47.831138 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 14:01:47 crc kubenswrapper[4715]: E1204 14:01:47.831333 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dhlrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nwvqj_openshift-marketplace(86f90c78-55bf-4118-b096-e9109678f1ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:01:47 crc kubenswrapper[4715]: E1204 14:01:47.832818 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nwvqj" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.441797 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.443195 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.452557 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.520091 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.520157 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kube-api-access\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.520183 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-var-lock\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.621206 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.621328 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.621391 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kube-api-access\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.621529 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-var-lock\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.621620 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-var-lock\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.638916 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kube-api-access\") pod \"installer-9-crc\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:49 crc kubenswrapper[4715]: I1204 14:01:49.773624 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.056943 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nwvqj" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.140814 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.140846 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.141319 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8x5hj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-m2rtt_openshift-marketplace(0b7a0295-de02-46f9-81f0-2dee1479af29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.141409 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z86nf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9kjvl_openshift-marketplace(314da00a-3748-4149-b36c-8ced9d9f3018): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.142488 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-m2rtt" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" Dec 04 14:01:50 crc kubenswrapper[4715]: E1204 14:01:50.142532 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9kjvl" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.167385 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-m2rtt" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.167539 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9kjvl" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.228453 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.228856 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hhbft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ssltf_openshift-marketplace(605514ff-63cc-4e34-a76e-44dd58be1c3e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.229831 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.229936 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lxmnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-46t2d_openshift-marketplace(53cf9350-c1c3-4557-8c23-2084355e0646): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.229957 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ssltf" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.231962 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-46t2d" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.254209 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.254372 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzhxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4tfcm_openshift-marketplace(e07d8a9f-8c02-4ac6-9e82-96564833487b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:01:51 crc kubenswrapper[4715]: E1204 14:01:51.255745 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4tfcm" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" Dec 04 14:01:53 crc kubenswrapper[4715]: E1204 14:01:53.993706 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-46t2d" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" Dec 04 14:01:53 crc kubenswrapper[4715]: E1204 14:01:53.993706 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4tfcm" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" Dec 04 14:01:53 crc kubenswrapper[4715]: E1204 14:01:53.993825 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2334343877/1\": happened during read: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 14:01:53 crc kubenswrapper[4715]: E1204 14:01:53.993868 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ssltf" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" Dec 04 14:01:53 crc kubenswrapper[4715]: E1204 14:01:53.994211 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78qvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vkbdv_openshift-marketplace(b1fb6e69-52de-41b9-acae-df242cab542e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2334343877/1\": happened during read: context canceled" logger="UnhandledError" Dec 04 14:01:53 crc kubenswrapper[4715]: E1204 14:01:53.995385 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage2334343877/1\\\": happened during read: context canceled\"" pod="openshift-marketplace/redhat-operators-vkbdv" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.391674 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 14:01:54 crc kubenswrapper[4715]: W1204 14:01:54.404172 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode478be46_943d_4933_8cbb_b6f577e7f164.slice/crio-c5f6dd5d2f64f56fe219b5620de69822e00cae3128442aaf35a2be70cbc1be87 WatchSource:0}: Error finding container c5f6dd5d2f64f56fe219b5620de69822e00cae3128442aaf35a2be70cbc1be87: Status 404 returned error can't find the container with id c5f6dd5d2f64f56fe219b5620de69822e00cae3128442aaf35a2be70cbc1be87 Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.489279 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 14:01:54 crc kubenswrapper[4715]: W1204 14:01:54.503456 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7c31b6e7_1c82_4ff4_822f_69c8aec95c43.slice/crio-033a44d5079077454c8566aefcdd3c7693b54f67dbbacec27e3095af74ba4e4f WatchSource:0}: Error finding container 033a44d5079077454c8566aefcdd3c7693b54f67dbbacec27e3095af74ba4e4f: Status 404 returned error can't find the container with id 033a44d5079077454c8566aefcdd3c7693b54f67dbbacec27e3095af74ba4e4f Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.897220 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e478be46-943d-4933-8cbb-b6f577e7f164","Type":"ContainerStarted","Data":"7a9a4d25f76bb9c80b7dcf1807191c0f0d2d8f7ccbd45d2f913b1c9f3bdfab08"} Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.897562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e478be46-943d-4933-8cbb-b6f577e7f164","Type":"ContainerStarted","Data":"c5f6dd5d2f64f56fe219b5620de69822e00cae3128442aaf35a2be70cbc1be87"} Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.899538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7c31b6e7-1c82-4ff4-822f-69c8aec95c43","Type":"ContainerStarted","Data":"815450724bfdab04b07c705052f8503157f4617ac33c9d8c49cb4463d43e0093"} Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.899570 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7c31b6e7-1c82-4ff4-822f-69c8aec95c43","Type":"ContainerStarted","Data":"033a44d5079077454c8566aefcdd3c7693b54f67dbbacec27e3095af74ba4e4f"} Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.901749 4715 generic.go:334] "Generic (PLEG): container finished" podID="8414683b-d211-4ea5-8927-f0252a960c72" containerID="094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817" exitCode=0 Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.901868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerDied","Data":"094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817"} Dec 04 14:01:54 crc kubenswrapper[4715]: E1204 14:01:54.905963 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vkbdv" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" Dec 04 14:01:54 crc kubenswrapper[4715]: I1204 14:01:54.918078 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=9.918059447 podStartE2EDuration="9.918059447s" podCreationTimestamp="2025-12-04 14:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:54.91560216 +0000 UTC m=+271.984320375" watchObservedRunningTime="2025-12-04 14:01:54.918059447 +0000 UTC m=+271.986777662" Dec 04 14:01:55 crc kubenswrapper[4715]: I1204 14:01:55.910694 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerStarted","Data":"cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d"} Dec 04 14:01:55 crc kubenswrapper[4715]: I1204 14:01:55.912484 4715 generic.go:334] "Generic (PLEG): container finished" podID="e478be46-943d-4933-8cbb-b6f577e7f164" containerID="7a9a4d25f76bb9c80b7dcf1807191c0f0d2d8f7ccbd45d2f913b1c9f3bdfab08" exitCode=0 Dec 04 14:01:55 crc kubenswrapper[4715]: I1204 14:01:55.913890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e478be46-943d-4933-8cbb-b6f577e7f164","Type":"ContainerDied","Data":"7a9a4d25f76bb9c80b7dcf1807191c0f0d2d8f7ccbd45d2f913b1c9f3bdfab08"} Dec 04 14:01:55 crc kubenswrapper[4715]: I1204 14:01:55.930406 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=6.9303890280000005 podStartE2EDuration="6.930389028s" podCreationTimestamp="2025-12-04 14:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:54.990721055 +0000 UTC m=+272.059439270" watchObservedRunningTime="2025-12-04 14:01:55.930389028 +0000 UTC m=+272.999107243" Dec 04 14:01:55 crc kubenswrapper[4715]: I1204 14:01:55.931888 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p8xqs" podStartSLOduration=5.286003579 podStartE2EDuration="45.931881369s" podCreationTimestamp="2025-12-04 14:01:10 +0000 UTC" firstStartedPulling="2025-12-04 14:01:14.664017095 +0000 UTC m=+231.732735310" lastFinishedPulling="2025-12-04 14:01:55.309894875 +0000 UTC m=+272.378613100" observedRunningTime="2025-12-04 14:01:55.930930453 +0000 UTC m=+272.999648658" watchObservedRunningTime="2025-12-04 14:01:55.931881369 +0000 UTC m=+273.000599584" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.712925 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.713005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.714864 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.715042 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.724482 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.728944 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.813906 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.814045 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.815777 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.826531 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.841583 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.842839 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.904908 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.934178 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 14:01:56 crc kubenswrapper[4715]: I1204 14:01:56.942313 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.155952 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.322471 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e478be46-943d-4933-8cbb-b6f577e7f164-kube-api-access\") pod \"e478be46-943d-4933-8cbb-b6f577e7f164\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.322627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e478be46-943d-4933-8cbb-b6f577e7f164-kubelet-dir\") pod \"e478be46-943d-4933-8cbb-b6f577e7f164\" (UID: \"e478be46-943d-4933-8cbb-b6f577e7f164\") " Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.322752 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e478be46-943d-4933-8cbb-b6f577e7f164-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e478be46-943d-4933-8cbb-b6f577e7f164" (UID: "e478be46-943d-4933-8cbb-b6f577e7f164"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.323366 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e478be46-943d-4933-8cbb-b6f577e7f164-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.328306 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e478be46-943d-4933-8cbb-b6f577e7f164-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e478be46-943d-4933-8cbb-b6f577e7f164" (UID: "e478be46-943d-4933-8cbb-b6f577e7f164"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:01:57 crc kubenswrapper[4715]: W1204 14:01:57.422741 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-df26d5d0cd687b6956cc2d81521028e33e265e2fce78fbc792cf10b4af71a2d5 WatchSource:0}: Error finding container df26d5d0cd687b6956cc2d81521028e33e265e2fce78fbc792cf10b4af71a2d5: Status 404 returned error can't find the container with id df26d5d0cd687b6956cc2d81521028e33e265e2fce78fbc792cf10b4af71a2d5 Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.424429 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e478be46-943d-4933-8cbb-b6f577e7f164-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.922631 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a84d0b807b0651de732501879f0a2e2934607e50b750d6a5fa671b7394f25b0b"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.922914 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"df26d5d0cd687b6956cc2d81521028e33e265e2fce78fbc792cf10b4af71a2d5"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.924883 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.924885 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e478be46-943d-4933-8cbb-b6f577e7f164","Type":"ContainerDied","Data":"c5f6dd5d2f64f56fe219b5620de69822e00cae3128442aaf35a2be70cbc1be87"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.924952 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5f6dd5d2f64f56fe219b5620de69822e00cae3128442aaf35a2be70cbc1be87" Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.926168 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"98a83b6172a17e19599c55792ed9740df4f407c90ae409fa3b1a7c0ae32c8962"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.926192 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"abccf7511392e56b27048e5a7b135767c0c6ff6503be53e1debb9c7b1c343134"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.927928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b2bf5cc1c1e1b59a5aad96d9e2cf7b5b5aa086edc6f8adb43acaa9165f45011d"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.927952 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c4dfcef12dbaa5c29bfea0a1c0642da01ca76c3ea4ab6aca4e47eeced74d6010"} Dec 04 14:01:57 crc kubenswrapper[4715]: I1204 14:01:57.928323 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:02:01 crc kubenswrapper[4715]: I1204 14:02:01.134529 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:02:01 crc kubenswrapper[4715]: I1204 14:02:01.134834 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:02:01 crc kubenswrapper[4715]: I1204 14:02:01.199693 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:02:01 crc kubenswrapper[4715]: I1204 14:02:01.994103 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:02:02 crc kubenswrapper[4715]: I1204 14:02:02.038800 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p8xqs"] Dec 04 14:02:03 crc kubenswrapper[4715]: I1204 14:02:03.956375 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p8xqs" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="registry-server" containerID="cri-o://cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d" gracePeriod=2 Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.768970 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.830917 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fntk\" (UniqueName: \"kubernetes.io/projected/8414683b-d211-4ea5-8927-f0252a960c72-kube-api-access-6fntk\") pod \"8414683b-d211-4ea5-8927-f0252a960c72\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.831944 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-utilities\") pod \"8414683b-d211-4ea5-8927-f0252a960c72\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.832022 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-catalog-content\") pod \"8414683b-d211-4ea5-8927-f0252a960c72\" (UID: \"8414683b-d211-4ea5-8927-f0252a960c72\") " Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.833879 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-utilities" (OuterVolumeSpecName: "utilities") pod "8414683b-d211-4ea5-8927-f0252a960c72" (UID: "8414683b-d211-4ea5-8927-f0252a960c72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.836356 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8414683b-d211-4ea5-8927-f0252a960c72-kube-api-access-6fntk" (OuterVolumeSpecName: "kube-api-access-6fntk") pod "8414683b-d211-4ea5-8927-f0252a960c72" (UID: "8414683b-d211-4ea5-8927-f0252a960c72"). InnerVolumeSpecName "kube-api-access-6fntk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.933469 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fntk\" (UniqueName: \"kubernetes.io/projected/8414683b-d211-4ea5-8927-f0252a960c72-kube-api-access-6fntk\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.933504 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.950622 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8414683b-d211-4ea5-8927-f0252a960c72" (UID: "8414683b-d211-4ea5-8927-f0252a960c72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.968134 4715 generic.go:334] "Generic (PLEG): container finished" podID="8414683b-d211-4ea5-8927-f0252a960c72" containerID="cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d" exitCode=0 Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.968218 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8xqs" Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.968208 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerDied","Data":"cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d"} Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.968376 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8xqs" event={"ID":"8414683b-d211-4ea5-8927-f0252a960c72","Type":"ContainerDied","Data":"257f93f61543427eaa8dfca9335020a92813a0d074123e1ecaa603e33f9f9780"} Dec 04 14:02:05 crc kubenswrapper[4715]: I1204 14:02:05.968429 4715 scope.go:117] "RemoveContainer" containerID="cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.001851 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p8xqs"] Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.004495 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p8xqs"] Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.034890 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8414683b-d211-4ea5-8927-f0252a960c72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.074535 4715 scope.go:117] "RemoveContainer" containerID="094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.091329 4715 scope.go:117] "RemoveContainer" containerID="2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.115261 4715 scope.go:117] "RemoveContainer" containerID="cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d" Dec 04 14:02:06 crc kubenswrapper[4715]: E1204 14:02:06.115766 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d\": container with ID starting with cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d not found: ID does not exist" containerID="cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.115822 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d"} err="failed to get container status \"cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d\": rpc error: code = NotFound desc = could not find container \"cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d\": container with ID starting with cbe51ab66e80ea97a6105c54762984fdc1d17d31ff52402d020969f06fa8186d not found: ID does not exist" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.115858 4715 scope.go:117] "RemoveContainer" containerID="094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817" Dec 04 14:02:06 crc kubenswrapper[4715]: E1204 14:02:06.116651 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817\": container with ID starting with 094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817 not found: ID does not exist" containerID="094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.116677 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817"} err="failed to get container status \"094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817\": rpc error: code = NotFound desc = could not find container \"094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817\": container with ID starting with 094e4ef4582fe21bdae2b8a949622d95c4be3d204b3a26bc21aae860cd2d3817 not found: ID does not exist" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.116695 4715 scope.go:117] "RemoveContainer" containerID="2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474" Dec 04 14:02:06 crc kubenswrapper[4715]: E1204 14:02:06.116938 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474\": container with ID starting with 2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474 not found: ID does not exist" containerID="2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.116962 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474"} err="failed to get container status \"2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474\": rpc error: code = NotFound desc = could not find container \"2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474\": container with ID starting with 2d41eefd2746c13ae6b2799b190a3a092a8e30d5c35aee7313f3d5fe5bbc2474 not found: ID does not exist" Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.975493 4715 generic.go:334] "Generic (PLEG): container finished" podID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerID="3548dafc6fdc5e567d123b45241830fbcf3e53ce2af8164388d8c9085092df8a" exitCode=0 Dec 04 14:02:06 crc kubenswrapper[4715]: I1204 14:02:06.975586 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2rtt" event={"ID":"0b7a0295-de02-46f9-81f0-2dee1479af29","Type":"ContainerDied","Data":"3548dafc6fdc5e567d123b45241830fbcf3e53ce2af8164388d8c9085092df8a"} Dec 04 14:02:07 crc kubenswrapper[4715]: I1204 14:02:07.192691 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8414683b-d211-4ea5-8927-f0252a960c72" path="/var/lib/kubelet/pods/8414683b-d211-4ea5-8927-f0252a960c72/volumes" Dec 04 14:02:08 crc kubenswrapper[4715]: I1204 14:02:08.644613 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nrpl2"] Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.018962 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerStarted","Data":"c92f3cc4c5cae56ef0d50a1306d6a54313a9bb4f333abbee440ebab60bd20bb5"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.033736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2rtt" event={"ID":"0b7a0295-de02-46f9-81f0-2dee1479af29","Type":"ContainerStarted","Data":"8ea09c624477d7e5f2fd524b8d3c1e764333085ce05977e635b82ccd02b2aa1c"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.044912 4715 generic.go:334] "Generic (PLEG): container finished" podID="86f90c78-55bf-4118-b096-e9109678f1ff" containerID="f9b928f65a2dca1f8b6dc33a74ae0fc457d57ca72f407070cc0a4a29e29ca8c5" exitCode=0 Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.044999 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwvqj" event={"ID":"86f90c78-55bf-4118-b096-e9109678f1ff","Type":"ContainerDied","Data":"f9b928f65a2dca1f8b6dc33a74ae0fc457d57ca72f407070cc0a4a29e29ca8c5"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.048344 4715 generic.go:334] "Generic (PLEG): container finished" podID="53cf9350-c1c3-4557-8c23-2084355e0646" containerID="ba78a344e984ef5333cc214efdf0f5a0af8d9f3f111cd7a6e4f48206e331b4f8" exitCode=0 Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.048395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46t2d" event={"ID":"53cf9350-c1c3-4557-8c23-2084355e0646","Type":"ContainerDied","Data":"ba78a344e984ef5333cc214efdf0f5a0af8d9f3f111cd7a6e4f48206e331b4f8"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.060332 4715 generic.go:334] "Generic (PLEG): container finished" podID="314da00a-3748-4149-b36c-8ced9d9f3018" containerID="12bcc374284a033ae37f3498707f3dbdc59beac350eca70c3081a725d32f2f74" exitCode=0 Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.060407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kjvl" event={"ID":"314da00a-3748-4149-b36c-8ced9d9f3018","Type":"ContainerDied","Data":"12bcc374284a033ae37f3498707f3dbdc59beac350eca70c3081a725d32f2f74"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.064155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerStarted","Data":"f8a868bf64774ddd3a6a751558977f4f90476ae571a71aaf5de87982f974db4a"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.070334 4715 generic.go:334] "Generic (PLEG): container finished" podID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerID="b66947568c6b3a10392adc037f526779e42a312bd38ae9ac7d91772db16ca09f" exitCode=0 Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.070382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ssltf" event={"ID":"605514ff-63cc-4e34-a76e-44dd58be1c3e","Type":"ContainerDied","Data":"b66947568c6b3a10392adc037f526779e42a312bd38ae9ac7d91772db16ca09f"} Dec 04 14:02:14 crc kubenswrapper[4715]: I1204 14:02:14.072109 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m2rtt" podStartSLOduration=3.778327034 podStartE2EDuration="1m7.072091353s" podCreationTimestamp="2025-12-04 14:01:07 +0000 UTC" firstStartedPulling="2025-12-04 14:01:09.428053273 +0000 UTC m=+226.496771488" lastFinishedPulling="2025-12-04 14:02:12.721817592 +0000 UTC m=+289.790535807" observedRunningTime="2025-12-04 14:02:14.070343946 +0000 UTC m=+291.139062171" watchObservedRunningTime="2025-12-04 14:02:14.072091353 +0000 UTC m=+291.140809568" Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.090948 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kjvl" event={"ID":"314da00a-3748-4149-b36c-8ced9d9f3018","Type":"ContainerStarted","Data":"c7d065177df835828ce208238792d767c634642b475ab21c8019215da0ad9c7a"} Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.094745 4715 generic.go:334] "Generic (PLEG): container finished" podID="b1fb6e69-52de-41b9-acae-df242cab542e" containerID="f8a868bf64774ddd3a6a751558977f4f90476ae571a71aaf5de87982f974db4a" exitCode=0 Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.094812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerDied","Data":"f8a868bf64774ddd3a6a751558977f4f90476ae571a71aaf5de87982f974db4a"} Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.099585 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ssltf" event={"ID":"605514ff-63cc-4e34-a76e-44dd58be1c3e","Type":"ContainerStarted","Data":"07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e"} Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.103710 4715 generic.go:334] "Generic (PLEG): container finished" podID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerID="c92f3cc4c5cae56ef0d50a1306d6a54313a9bb4f333abbee440ebab60bd20bb5" exitCode=0 Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.103775 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerDied","Data":"c92f3cc4c5cae56ef0d50a1306d6a54313a9bb4f333abbee440ebab60bd20bb5"} Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.108877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwvqj" event={"ID":"86f90c78-55bf-4118-b096-e9109678f1ff","Type":"ContainerStarted","Data":"6e96211274c685acadb58b12466bcf6e6dfe078ee943721477454a036737cd34"} Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.113509 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9kjvl" podStartSLOduration=2.968129867 podStartE2EDuration="1m8.113494486s" podCreationTimestamp="2025-12-04 14:01:07 +0000 UTC" firstStartedPulling="2025-12-04 14:01:09.428544727 +0000 UTC m=+226.497262942" lastFinishedPulling="2025-12-04 14:02:14.573909346 +0000 UTC m=+291.642627561" observedRunningTime="2025-12-04 14:02:15.107720919 +0000 UTC m=+292.176439144" watchObservedRunningTime="2025-12-04 14:02:15.113494486 +0000 UTC m=+292.182212701" Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.114387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46t2d" event={"ID":"53cf9350-c1c3-4557-8c23-2084355e0646","Type":"ContainerStarted","Data":"c1a7874a99acb7b812f946c57cc4b4ea125786a2109850dc5cf92a89d130ff60"} Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.171211 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ssltf" podStartSLOduration=2.152569878 podStartE2EDuration="1m6.171192786s" podCreationTimestamp="2025-12-04 14:01:09 +0000 UTC" firstStartedPulling="2025-12-04 14:01:10.478638411 +0000 UTC m=+227.547356626" lastFinishedPulling="2025-12-04 14:02:14.497261299 +0000 UTC m=+291.565979534" observedRunningTime="2025-12-04 14:02:15.1661939 +0000 UTC m=+292.234912115" watchObservedRunningTime="2025-12-04 14:02:15.171192786 +0000 UTC m=+292.239911001" Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.215712 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwvqj" podStartSLOduration=1.902127229 podStartE2EDuration="1m8.215662707s" podCreationTimestamp="2025-12-04 14:01:07 +0000 UTC" firstStartedPulling="2025-12-04 14:01:08.388841276 +0000 UTC m=+225.457559491" lastFinishedPulling="2025-12-04 14:02:14.702376754 +0000 UTC m=+291.771094969" observedRunningTime="2025-12-04 14:02:15.208235064 +0000 UTC m=+292.276953279" watchObservedRunningTime="2025-12-04 14:02:15.215662707 +0000 UTC m=+292.284380942" Dec 04 14:02:15 crc kubenswrapper[4715]: I1204 14:02:15.227437 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-46t2d" podStartSLOduration=3.265973543 podStartE2EDuration="1m6.227417097s" podCreationTimestamp="2025-12-04 14:01:09 +0000 UTC" firstStartedPulling="2025-12-04 14:01:11.506313542 +0000 UTC m=+228.575031757" lastFinishedPulling="2025-12-04 14:02:14.467757096 +0000 UTC m=+291.536475311" observedRunningTime="2025-12-04 14:02:15.225605247 +0000 UTC m=+292.294323482" watchObservedRunningTime="2025-12-04 14:02:15.227417097 +0000 UTC m=+292.296135312" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.010426 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.013381 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.026486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4cf68f32-a8c0-41c4-a53d-5f3dbd01197e-metrics-certs\") pod \"network-metrics-daemon-c9ttc\" (UID: \"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e\") " pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.125166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.128362 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerStarted","Data":"0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6"} Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.132009 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c9ttc" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.178340 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vkbdv" podStartSLOduration=3.169165099 podStartE2EDuration="1m6.178325575s" podCreationTimestamp="2025-12-04 14:01:10 +0000 UTC" firstStartedPulling="2025-12-04 14:01:12.567631192 +0000 UTC m=+229.636349407" lastFinishedPulling="2025-12-04 14:02:15.576791678 +0000 UTC m=+292.645509883" observedRunningTime="2025-12-04 14:02:16.176388352 +0000 UTC m=+293.245106567" watchObservedRunningTime="2025-12-04 14:02:16.178325575 +0000 UTC m=+293.247043790" Dec 04 14:02:16 crc kubenswrapper[4715]: I1204 14:02:16.460319 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-c9ttc"] Dec 04 14:02:16 crc kubenswrapper[4715]: W1204 14:02:16.474970 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cf68f32_a8c0_41c4_a53d_5f3dbd01197e.slice/crio-3836bb05ad99e8b8f862db42afe97ae8afd1384af496a7c6c44fbc96e0e5b26a WatchSource:0}: Error finding container 3836bb05ad99e8b8f862db42afe97ae8afd1384af496a7c6c44fbc96e0e5b26a: Status 404 returned error can't find the container with id 3836bb05ad99e8b8f862db42afe97ae8afd1384af496a7c6c44fbc96e0e5b26a Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.137428 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" event={"ID":"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e","Type":"ContainerStarted","Data":"f028192f50c2baba01ff20e834210cb13a5929d440a6bf4e66ca71972c9ed9a2"} Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.137769 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" event={"ID":"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e","Type":"ContainerStarted","Data":"d8cd9fea17ec2a2e0f2bf69745da75696dcdd7d216b4565f6a6f3382f2a46241"} Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.137783 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c9ttc" event={"ID":"4cf68f32-a8c0-41c4-a53d-5f3dbd01197e","Type":"ContainerStarted","Data":"3836bb05ad99e8b8f862db42afe97ae8afd1384af496a7c6c44fbc96e0e5b26a"} Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.155488 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-c9ttc" podStartSLOduration=266.155471058 podStartE2EDuration="4m26.155471058s" podCreationTimestamp="2025-12-04 13:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:02:17.151199412 +0000 UTC m=+294.219917627" watchObservedRunningTime="2025-12-04 14:02:17.155471058 +0000 UTC m=+294.224189273" Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.683757 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.684092 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.738712 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.905990 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.906069 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:02:17 crc kubenswrapper[4715]: I1204 14:02:17.958803 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:02:18 crc kubenswrapper[4715]: I1204 14:02:18.121512 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:02:18 crc kubenswrapper[4715]: I1204 14:02:18.121580 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:02:18 crc kubenswrapper[4715]: I1204 14:02:18.144963 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerStarted","Data":"445542630ea202f186bd7cc4b877cdad1310c145bbbdac9eb61623c183ace444"} Dec 04 14:02:18 crc kubenswrapper[4715]: I1204 14:02:18.169813 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4tfcm" podStartSLOduration=2.198879721 podStartE2EDuration="1m11.169787963s" podCreationTimestamp="2025-12-04 14:01:07 +0000 UTC" firstStartedPulling="2025-12-04 14:01:08.406432626 +0000 UTC m=+225.475150841" lastFinishedPulling="2025-12-04 14:02:17.377340858 +0000 UTC m=+294.446059083" observedRunningTime="2025-12-04 14:02:18.165732692 +0000 UTC m=+295.234450917" watchObservedRunningTime="2025-12-04 14:02:18.169787963 +0000 UTC m=+295.238506198" Dec 04 14:02:18 crc kubenswrapper[4715]: I1204 14:02:18.189006 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:02:18 crc kubenswrapper[4715]: I1204 14:02:18.256736 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:02:19 crc kubenswrapper[4715]: I1204 14:02:19.704763 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:02:19 crc kubenswrapper[4715]: I1204 14:02:19.704825 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.057490 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.122602 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.123236 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.165996 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.199401 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.207588 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.433364 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m2rtt"] Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.434061 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m2rtt" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="registry-server" containerID="cri-o://8ea09c624477d7e5f2fd524b8d3c1e764333085ce05977e635b82ccd02b2aa1c" gracePeriod=2 Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.766754 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:02:20 crc kubenswrapper[4715]: I1204 14:02:20.767411 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:02:21 crc kubenswrapper[4715]: I1204 14:02:21.805795 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vkbdv" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="registry-server" probeResult="failure" output=< Dec 04 14:02:21 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 04 14:02:21 crc kubenswrapper[4715]: > Dec 04 14:02:23 crc kubenswrapper[4715]: I1204 14:02:23.434992 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46t2d"] Dec 04 14:02:23 crc kubenswrapper[4715]: I1204 14:02:23.435227 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-46t2d" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="registry-server" containerID="cri-o://c1a7874a99acb7b812f946c57cc4b4ea125786a2109850dc5cf92a89d130ff60" gracePeriod=2 Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.303129 4715 generic.go:334] "Generic (PLEG): container finished" podID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerID="8ea09c624477d7e5f2fd524b8d3c1e764333085ce05977e635b82ccd02b2aa1c" exitCode=0 Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.303583 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2rtt" event={"ID":"0b7a0295-de02-46f9-81f0-2dee1479af29","Type":"ContainerDied","Data":"8ea09c624477d7e5f2fd524b8d3c1e764333085ce05977e635b82ccd02b2aa1c"} Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.407563 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.556789 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-utilities\") pod \"0b7a0295-de02-46f9-81f0-2dee1479af29\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.556915 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x5hj\" (UniqueName: \"kubernetes.io/projected/0b7a0295-de02-46f9-81f0-2dee1479af29-kube-api-access-8x5hj\") pod \"0b7a0295-de02-46f9-81f0-2dee1479af29\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.556955 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-catalog-content\") pod \"0b7a0295-de02-46f9-81f0-2dee1479af29\" (UID: \"0b7a0295-de02-46f9-81f0-2dee1479af29\") " Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.558002 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-utilities" (OuterVolumeSpecName: "utilities") pod "0b7a0295-de02-46f9-81f0-2dee1479af29" (UID: "0b7a0295-de02-46f9-81f0-2dee1479af29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.562688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7a0295-de02-46f9-81f0-2dee1479af29-kube-api-access-8x5hj" (OuterVolumeSpecName: "kube-api-access-8x5hj") pod "0b7a0295-de02-46f9-81f0-2dee1479af29" (UID: "0b7a0295-de02-46f9-81f0-2dee1479af29"). InnerVolumeSpecName "kube-api-access-8x5hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.604494 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b7a0295-de02-46f9-81f0-2dee1479af29" (UID: "0b7a0295-de02-46f9-81f0-2dee1479af29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.657962 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.658009 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x5hj\" (UniqueName: \"kubernetes.io/projected/0b7a0295-de02-46f9-81f0-2dee1479af29-kube-api-access-8x5hj\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:26 crc kubenswrapper[4715]: I1204 14:02:26.658020 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b7a0295-de02-46f9-81f0-2dee1479af29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.311233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2rtt" event={"ID":"0b7a0295-de02-46f9-81f0-2dee1479af29","Type":"ContainerDied","Data":"79db332a5c33462b2ebcd9485ff427d39af6c90570b12f591620550a93116176"} Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.311277 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2rtt" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.311297 4715 scope.go:117] "RemoveContainer" containerID="8ea09c624477d7e5f2fd524b8d3c1e764333085ce05977e635b82ccd02b2aa1c" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.317891 4715 generic.go:334] "Generic (PLEG): container finished" podID="53cf9350-c1c3-4557-8c23-2084355e0646" containerID="c1a7874a99acb7b812f946c57cc4b4ea125786a2109850dc5cf92a89d130ff60" exitCode=0 Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.317954 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46t2d" event={"ID":"53cf9350-c1c3-4557-8c23-2084355e0646","Type":"ContainerDied","Data":"c1a7874a99acb7b812f946c57cc4b4ea125786a2109850dc5cf92a89d130ff60"} Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.341427 4715 scope.go:117] "RemoveContainer" containerID="3548dafc6fdc5e567d123b45241830fbcf3e53ce2af8164388d8c9085092df8a" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.342468 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m2rtt"] Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.346139 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m2rtt"] Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.358366 4715 scope.go:117] "RemoveContainer" containerID="d2d984631c7977b509167cde50585897c8013555a9027046432308b7a8e4f8b6" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.493605 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.493684 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.529902 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.719245 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:02:27 crc kubenswrapper[4715]: I1204 14:02:27.964687 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:02:28 crc kubenswrapper[4715]: I1204 14:02:28.362980 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.190186 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" path="/var/lib/kubelet/pods/0b7a0295-de02-46f9-81f0-2dee1479af29/volumes" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.609598 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.803617 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-utilities\") pod \"53cf9350-c1c3-4557-8c23-2084355e0646\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.803678 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxmnd\" (UniqueName: \"kubernetes.io/projected/53cf9350-c1c3-4557-8c23-2084355e0646-kube-api-access-lxmnd\") pod \"53cf9350-c1c3-4557-8c23-2084355e0646\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.803723 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-catalog-content\") pod \"53cf9350-c1c3-4557-8c23-2084355e0646\" (UID: \"53cf9350-c1c3-4557-8c23-2084355e0646\") " Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.804688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-utilities" (OuterVolumeSpecName: "utilities") pod "53cf9350-c1c3-4557-8c23-2084355e0646" (UID: "53cf9350-c1c3-4557-8c23-2084355e0646"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.810203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53cf9350-c1c3-4557-8c23-2084355e0646-kube-api-access-lxmnd" (OuterVolumeSpecName: "kube-api-access-lxmnd") pod "53cf9350-c1c3-4557-8c23-2084355e0646" (UID: "53cf9350-c1c3-4557-8c23-2084355e0646"). InnerVolumeSpecName "kube-api-access-lxmnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.822929 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53cf9350-c1c3-4557-8c23-2084355e0646" (UID: "53cf9350-c1c3-4557-8c23-2084355e0646"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.904933 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.904982 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxmnd\" (UniqueName: \"kubernetes.io/projected/53cf9350-c1c3-4557-8c23-2084355e0646-kube-api-access-lxmnd\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:29 crc kubenswrapper[4715]: I1204 14:02:29.904996 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53cf9350-c1c3-4557-8c23-2084355e0646-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.338156 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46t2d" event={"ID":"53cf9350-c1c3-4557-8c23-2084355e0646","Type":"ContainerDied","Data":"c7b64722f3dbb7cfd6dc4f774071f34c0a3c0c02cd57927dc4404944684bb320"} Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.338214 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46t2d" Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.338223 4715 scope.go:117] "RemoveContainer" containerID="c1a7874a99acb7b812f946c57cc4b4ea125786a2109850dc5cf92a89d130ff60" Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.355481 4715 scope.go:117] "RemoveContainer" containerID="ba78a344e984ef5333cc214efdf0f5a0af8d9f3f111cd7a6e4f48206e331b4f8" Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.374345 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46t2d"] Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.375641 4715 scope.go:117] "RemoveContainer" containerID="ca17907fe2b8cc4e1c47f5217874b0a7a0dde71fb7a9434f6322f764719f52db" Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.378288 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-46t2d"] Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.640047 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwvqj"] Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.640281 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwvqj" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="registry-server" containerID="cri-o://6e96211274c685acadb58b12466bcf6e6dfe078ee943721477454a036737cd34" gracePeriod=2 Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.810944 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:02:30 crc kubenswrapper[4715]: I1204 14:02:30.853985 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.189516 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" path="/var/lib/kubelet/pods/53cf9350-c1c3-4557-8c23-2084355e0646/volumes" Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.345876 4715 generic.go:334] "Generic (PLEG): container finished" podID="86f90c78-55bf-4118-b096-e9109678f1ff" containerID="6e96211274c685acadb58b12466bcf6e6dfe078ee943721477454a036737cd34" exitCode=0 Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.345933 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwvqj" event={"ID":"86f90c78-55bf-4118-b096-e9109678f1ff","Type":"ContainerDied","Data":"6e96211274c685acadb58b12466bcf6e6dfe078ee943721477454a036737cd34"} Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.802723 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.932736 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-catalog-content\") pod \"86f90c78-55bf-4118-b096-e9109678f1ff\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.932904 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhlrl\" (UniqueName: \"kubernetes.io/projected/86f90c78-55bf-4118-b096-e9109678f1ff-kube-api-access-dhlrl\") pod \"86f90c78-55bf-4118-b096-e9109678f1ff\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.932936 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-utilities\") pod \"86f90c78-55bf-4118-b096-e9109678f1ff\" (UID: \"86f90c78-55bf-4118-b096-e9109678f1ff\") " Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.933761 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-utilities" (OuterVolumeSpecName: "utilities") pod "86f90c78-55bf-4118-b096-e9109678f1ff" (UID: "86f90c78-55bf-4118-b096-e9109678f1ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.938272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f90c78-55bf-4118-b096-e9109678f1ff-kube-api-access-dhlrl" (OuterVolumeSpecName: "kube-api-access-dhlrl") pod "86f90c78-55bf-4118-b096-e9109678f1ff" (UID: "86f90c78-55bf-4118-b096-e9109678f1ff"). InnerVolumeSpecName "kube-api-access-dhlrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:31 crc kubenswrapper[4715]: I1204 14:02:31.979822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86f90c78-55bf-4118-b096-e9109678f1ff" (UID: "86f90c78-55bf-4118-b096-e9109678f1ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.034225 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.034267 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhlrl\" (UniqueName: \"kubernetes.io/projected/86f90c78-55bf-4118-b096-e9109678f1ff-kube-api-access-dhlrl\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.034281 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f90c78-55bf-4118-b096-e9109678f1ff-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.354865 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwvqj" event={"ID":"86f90c78-55bf-4118-b096-e9109678f1ff","Type":"ContainerDied","Data":"97a6d27b2cc37a54c4b093fbda0ded5c7382c31b4a1ea798ce6159fa69b3d7d4"} Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.354931 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwvqj" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.354933 4715 scope.go:117] "RemoveContainer" containerID="6e96211274c685acadb58b12466bcf6e6dfe078ee943721477454a036737cd34" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.381427 4715 scope.go:117] "RemoveContainer" containerID="f9b928f65a2dca1f8b6dc33a74ae0fc457d57ca72f407070cc0a4a29e29ca8c5" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.386617 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwvqj"] Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.390288 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwvqj"] Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.397698 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.397981 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.397995 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398007 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398014 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398044 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e478be46-943d-4933-8cbb-b6f577e7f164" containerName="pruner" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398051 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e478be46-943d-4933-8cbb-b6f577e7f164" containerName="pruner" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398058 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398065 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398077 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398084 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398094 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398102 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398111 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398117 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398129 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398135 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398144 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398151 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398163 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398170 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398180 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398186 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="extract-utilities" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398196 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398204 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.398214 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398221 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="extract-content" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398325 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7a0295-de02-46f9-81f0-2dee1479af29" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398337 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e478be46-943d-4933-8cbb-b6f577e7f164" containerName="pruner" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398350 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="53cf9350-c1c3-4557-8c23-2084355e0646" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398360 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8414683b-d211-4ea5-8927-f0252a960c72" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398368 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" containerName="registry-server" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.398915 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.401047 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.401431 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55" gracePeriod=15 Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.401449 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c" gracePeriod=15 Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.401564 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399" gracePeriod=15 Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.401625 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96" gracePeriod=15 Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.401664 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c" gracePeriod=15 Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.403868 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.404206 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404225 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.404238 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404245 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.404273 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404281 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.404291 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404297 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.404311 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404320 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.404359 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404366 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404528 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404541 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404550 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404559 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.404587 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.420190 4715 scope.go:117] "RemoveContainer" containerID="37fc4be1d4d228abfc0721b4f0f8ce3133caf583c354f25c71be698f81b6a6d3" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442663 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442722 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442743 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442763 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442925 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.442945 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.482129 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543620 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543673 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543684 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543712 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543740 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543767 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543777 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543756 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543793 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543824 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543789 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.543828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.750520 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.751080 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.751434 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.751740 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.752077 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.752124 4715 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.752456 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="200ms" Dec 04 14:02:32 crc kubenswrapper[4715]: I1204 14:02:32.783673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:32 crc kubenswrapper[4715]: W1204 14:02:32.815786 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-82446e8febda6aff97ad92852e1f1bb513d978c1cf49b6f1e237e5d5611da494 WatchSource:0}: Error finding container 82446e8febda6aff97ad92852e1f1bb513d978c1cf49b6f1e237e5d5611da494: Status 404 returned error can't find the container with id 82446e8febda6aff97ad92852e1f1bb513d978c1cf49b6f1e237e5d5611da494 Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.819679 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e080c4825f51c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 14:02:32.818890012 +0000 UTC m=+309.887608227,LastTimestamp:2025-12-04 14:02:32.818890012 +0000 UTC m=+309.887608227,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 14:02:32 crc kubenswrapper[4715]: E1204 14:02:32.953310 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="400ms" Dec 04 14:02:33 crc kubenswrapper[4715]: I1204 14:02:33.183341 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:33 crc kubenswrapper[4715]: I1204 14:02:33.186983 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86f90c78-55bf-4118-b096-e9109678f1ff" path="/var/lib/kubelet/pods/86f90c78-55bf-4118-b096-e9109678f1ff/volumes" Dec 04 14:02:33 crc kubenswrapper[4715]: E1204 14:02:33.354152 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="800ms" Dec 04 14:02:33 crc kubenswrapper[4715]: I1204 14:02:33.360907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"82446e8febda6aff97ad92852e1f1bb513d978c1cf49b6f1e237e5d5611da494"} Dec 04 14:02:33 crc kubenswrapper[4715]: I1204 14:02:33.676641 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" containerName="oauth-openshift" containerID="cri-o://46bf7866d106c6933fa0e88e1b6be150124d0ce6e9de73aa69db3c3cba96a76c" gracePeriod=15 Dec 04 14:02:34 crc kubenswrapper[4715]: E1204 14:02:34.155673 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="1.6s" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.378566 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" containerID="815450724bfdab04b07c705052f8503157f4617ac33c9d8c49cb4463d43e0093" exitCode=0 Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.378680 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7c31b6e7-1c82-4ff4-822f-69c8aec95c43","Type":"ContainerDied","Data":"815450724bfdab04b07c705052f8503157f4617ac33c9d8c49cb4463d43e0093"} Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.379851 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.380772 4715 generic.go:334] "Generic (PLEG): container finished" podID="3f118cab-ae82-49c9-90fb-e00688b70c5b" containerID="46bf7866d106c6933fa0e88e1b6be150124d0ce6e9de73aa69db3c3cba96a76c" exitCode=0 Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.380870 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" event={"ID":"3f118cab-ae82-49c9-90fb-e00688b70c5b","Type":"ContainerDied","Data":"46bf7866d106c6933fa0e88e1b6be150124d0ce6e9de73aa69db3c3cba96a76c"} Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.384777 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.385602 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c" exitCode=0 Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.385643 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96" exitCode=0 Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.385660 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c" exitCode=0 Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.385669 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399" exitCode=2 Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.387331 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b3ae584922f448138f2eefee3b1f33bcd0dd1d8df86851152bbe70e4dfc42c4b"} Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.802417 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.803202 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.803431 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.998852 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-service-ca\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.998999 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-policies\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999025 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-cliconfig\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999117 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-error\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999138 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wj72\" (UniqueName: \"kubernetes.io/projected/3f118cab-ae82-49c9-90fb-e00688b70c5b-kube-api-access-2wj72\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999158 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-dir\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999189 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-ocp-branding-template\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999229 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-login\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999299 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-idp-0-file-data\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999345 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-session\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999390 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-router-certs\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999383 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999416 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-serving-cert\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999452 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-trusted-ca-bundle\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999485 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-provider-selection\") pod \"3f118cab-ae82-49c9-90fb-e00688b70c5b\" (UID: \"3f118cab-ae82-49c9-90fb-e00688b70c5b\") " Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999664 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999823 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:02:34 crc kubenswrapper[4715]: I1204 14:02:34.999935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.000199 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.000928 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.007507 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.008105 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.008674 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.008773 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f118cab-ae82-49c9-90fb-e00688b70c5b-kube-api-access-2wj72" (OuterVolumeSpecName: "kube-api-access-2wj72") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "kube-api-access-2wj72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.009590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.009649 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.009777 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.010545 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.010673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3f118cab-ae82-49c9-90fb-e00688b70c5b" (UID: "3f118cab-ae82-49c9-90fb-e00688b70c5b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101519 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wj72\" (UniqueName: \"kubernetes.io/projected/3f118cab-ae82-49c9-90fb-e00688b70c5b-kube-api-access-2wj72\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101561 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101577 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101591 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101603 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101618 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101630 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101643 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101659 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101674 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101687 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101705 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.101743 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f118cab-ae82-49c9-90fb-e00688b70c5b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.276524 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.277410 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.278401 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.279143 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.279442 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.398493 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.399314 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55" exitCode=0 Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.399387 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.399397 4715 scope.go:117] "RemoveContainer" containerID="e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.402254 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.402244 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" event={"ID":"3f118cab-ae82-49c9-90fb-e00688b70c5b","Type":"ContainerDied","Data":"b0fb0ad986bdb9309c4210e041607c67ba279b4a35c699341dd4386bf7b11d97"} Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.404589 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.404667 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.404929 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.405261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.405298 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.405706 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.405814 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.405018 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.406212 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.406986 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.407346 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.407814 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.408182 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.409186 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.409778 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.411193 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.424207 4715 scope.go:117] "RemoveContainer" containerID="14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.447973 4715 scope.go:117] "RemoveContainer" containerID="2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.482428 4715 scope.go:117] "RemoveContainer" containerID="5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.506125 4715 scope.go:117] "RemoveContainer" containerID="3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.506549 4715 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.506593 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.506758 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.521855 4715 scope.go:117] "RemoveContainer" containerID="aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.547990 4715 scope.go:117] "RemoveContainer" containerID="e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.549082 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\": container with ID starting with e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c not found: ID does not exist" containerID="e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.549119 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c"} err="failed to get container status \"e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\": rpc error: code = NotFound desc = could not find container \"e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c\": container with ID starting with e4cb9f0fb73e71113c28c8e3e4130bf73b5117e06ddf7213c4fb28282babc78c not found: ID does not exist" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.549140 4715 scope.go:117] "RemoveContainer" containerID="14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.549562 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\": container with ID starting with 14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96 not found: ID does not exist" containerID="14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.549605 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96"} err="failed to get container status \"14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\": rpc error: code = NotFound desc = could not find container \"14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96\": container with ID starting with 14636460320e82b6282e15e1a66a53ec5567e9f0806762f1bf580c135e687e96 not found: ID does not exist" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.549635 4715 scope.go:117] "RemoveContainer" containerID="2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.549939 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\": container with ID starting with 2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c not found: ID does not exist" containerID="2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.549967 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c"} err="failed to get container status \"2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\": rpc error: code = NotFound desc = could not find container \"2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c\": container with ID starting with 2ec4d743db8eb3169ed08c250b531d613dbdcdb1018d7b712a58c6b7fa09333c not found: ID does not exist" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.550553 4715 scope.go:117] "RemoveContainer" containerID="5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.550933 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\": container with ID starting with 5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399 not found: ID does not exist" containerID="5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.550960 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399"} err="failed to get container status \"5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\": rpc error: code = NotFound desc = could not find container \"5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399\": container with ID starting with 5d3fb13343031ee9bff9002406aa8ff407cf9e00c9989dd3f5fdf7e7360f2399 not found: ID does not exist" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.550976 4715 scope.go:117] "RemoveContainer" containerID="3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.551346 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\": container with ID starting with 3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55 not found: ID does not exist" containerID="3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.551385 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55"} err="failed to get container status \"3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\": rpc error: code = NotFound desc = could not find container \"3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55\": container with ID starting with 3232198cd6c5cb6802e54bd3391c64baf4c66b84c1e75d0896ca1fb6f58cec55 not found: ID does not exist" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.551417 4715 scope.go:117] "RemoveContainer" containerID="aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.551975 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\": container with ID starting with aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe not found: ID does not exist" containerID="aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.552339 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe"} err="failed to get container status \"aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\": rpc error: code = NotFound desc = could not find container \"aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe\": container with ID starting with aeef6b7c783911bf3035c9f0f1ebb3a7997354a78ea3c936cb8935b5c24617fe not found: ID does not exist" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.552353 4715 scope.go:117] "RemoveContainer" containerID="46bf7866d106c6933fa0e88e1b6be150124d0ce6e9de73aa69db3c3cba96a76c" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.624701 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.625530 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.628163 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.628649 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.715863 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.716351 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.716836 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:35 crc kubenswrapper[4715]: E1204 14:02:35.757188 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="3.2s" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.809789 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kubelet-dir\") pod \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.809894 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7c31b6e7-1c82-4ff4-822f-69c8aec95c43" (UID: "7c31b6e7-1c82-4ff4-822f-69c8aec95c43"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.810017 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kube-api-access\") pod \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.810108 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-var-lock\") pod \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\" (UID: \"7c31b6e7-1c82-4ff4-822f-69c8aec95c43\") " Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.810190 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-var-lock" (OuterVolumeSpecName: "var-lock") pod "7c31b6e7-1c82-4ff4-822f-69c8aec95c43" (UID: "7c31b6e7-1c82-4ff4-822f-69c8aec95c43"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.810713 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.810785 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.819507 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7c31b6e7-1c82-4ff4-822f-69c8aec95c43" (UID: "7c31b6e7-1c82-4ff4-822f-69c8aec95c43"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4715]: I1204 14:02:35.911246 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c31b6e7-1c82-4ff4-822f-69c8aec95c43-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.409656 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7c31b6e7-1c82-4ff4-822f-69c8aec95c43","Type":"ContainerDied","Data":"033a44d5079077454c8566aefcdd3c7693b54f67dbbacec27e3095af74ba4e4f"} Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.409733 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.409750 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="033a44d5079077454c8566aefcdd3c7693b54f67dbbacec27e3095af74ba4e4f" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.424568 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.425088 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.426010 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.910693 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.911494 4715 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.911940 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.912291 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:36 crc kubenswrapper[4715]: I1204 14:02:36.912604 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:37 crc kubenswrapper[4715]: I1204 14:02:37.188093 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 14:02:38 crc kubenswrapper[4715]: E1204 14:02:38.693874 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e080c4825f51c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 14:02:32.818890012 +0000 UTC m=+309.887608227,LastTimestamp:2025-12-04 14:02:32.818890012 +0000 UTC m=+309.887608227,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 14:02:38 crc kubenswrapper[4715]: E1204 14:02:38.958082 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="6.4s" Dec 04 14:02:43 crc kubenswrapper[4715]: I1204 14:02:43.185157 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:43 crc kubenswrapper[4715]: I1204 14:02:43.186111 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:43 crc kubenswrapper[4715]: I1204 14:02:43.186785 4715 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:45 crc kubenswrapper[4715]: E1204 14:02:45.359379 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="7s" Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.474209 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.474267 4715 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496" exitCode=1 Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.474298 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496"} Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.474755 4715 scope.go:117] "RemoveContainer" containerID="5b51c91ad1d5313deec494adbe3799bc34bc2361d94d5ef863a575e47295c496" Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.475735 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.476054 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.477361 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:46 crc kubenswrapper[4715]: I1204 14:02:46.477628 4715 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.180188 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.181988 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.182614 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.185747 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.186191 4715 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.193994 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.194027 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:47 crc kubenswrapper[4715]: E1204 14:02:47.194717 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.195275 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.344013 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.481749 4715 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="009a26751b590c935ddec40c3c1c5bc9a65a014b257ebe866925b20eebcc1dda" exitCode=0 Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.481822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"009a26751b590c935ddec40c3c1c5bc9a65a014b257ebe866925b20eebcc1dda"} Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.481850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ece89bc09babd116149b834d8c4e8ce614b0a47b5abb38ea2b5f2e32e5a95aa9"} Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.482117 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.482130 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:47 crc kubenswrapper[4715]: E1204 14:02:47.482562 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.482764 4715 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.482943 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.483174 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.484272 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.487291 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.487340 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"128aede1c7f266f80eb9d958c68096e30307e2b1ba64235d7543ebe098d2991c"} Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.488111 4715 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.488537 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.488918 4715 status_manager.go:851] "Failed to get status for pod" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:47 crc kubenswrapper[4715]: I1204 14:02:47.489371 4715 status_manager.go:851] "Failed to get status for pod" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" pod="openshift-authentication/oauth-openshift-558db77b4-nrpl2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nrpl2\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 04 14:02:48 crc kubenswrapper[4715]: I1204 14:02:48.494608 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0b3c1ad5197e6afcfbd518bcd529968335c8129d02d6b839b632f046cbf3e324"} Dec 04 14:02:48 crc kubenswrapper[4715]: I1204 14:02:48.494907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"44f6ced4ff636143c60b7f89f24f560c55f399247d7c0dca8e7f4dcc3ff416f0"} Dec 04 14:02:49 crc kubenswrapper[4715]: I1204 14:02:49.502627 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"327f2c3552fde1a5ccb4daf38c3e800718159fc86ee0a78c04ec25360b293df8"} Dec 04 14:02:50 crc kubenswrapper[4715]: I1204 14:02:50.512191 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"194d918f4825374b7027c02eb02ff69421be10a7544b19f2d01533dc7f2f8803"} Dec 04 14:02:51 crc kubenswrapper[4715]: I1204 14:02:51.520507 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fb6ed5a02b628246a37ee33a02bd52f8a078e95948e324c9088b0ffa450f3d0f"} Dec 04 14:02:51 crc kubenswrapper[4715]: I1204 14:02:51.520812 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:51 crc kubenswrapper[4715]: I1204 14:02:51.520842 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:51 crc kubenswrapper[4715]: I1204 14:02:51.520915 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:51 crc kubenswrapper[4715]: I1204 14:02:51.530520 4715 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.037057 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.195403 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.195453 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.201266 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.525226 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.525258 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:52 crc kubenswrapper[4715]: I1204 14:02:52.529832 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:02:53 crc kubenswrapper[4715]: I1204 14:02:53.199267 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ded298be-8ce0-485d-8de3-42e901c11993" Dec 04 14:02:53 crc kubenswrapper[4715]: I1204 14:02:53.529779 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:53 crc kubenswrapper[4715]: I1204 14:02:53.529815 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:53 crc kubenswrapper[4715]: I1204 14:02:53.533447 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ded298be-8ce0-485d-8de3-42e901c11993" Dec 04 14:02:54 crc kubenswrapper[4715]: I1204 14:02:54.535268 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:54 crc kubenswrapper[4715]: I1204 14:02:54.535296 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a754c83-0416-4a03-95c4-59a98329e921" Dec 04 14:02:54 crc kubenswrapper[4715]: I1204 14:02:54.538425 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ded298be-8ce0-485d-8de3-42e901c11993" Dec 04 14:02:54 crc kubenswrapper[4715]: I1204 14:02:54.861713 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 14:02:54 crc kubenswrapper[4715]: I1204 14:02:54.865323 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 14:03:02 crc kubenswrapper[4715]: I1204 14:03:02.040970 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 14:03:03 crc kubenswrapper[4715]: I1204 14:03:03.460673 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 14:03:04 crc kubenswrapper[4715]: I1204 14:03:04.468814 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 14:03:04 crc kubenswrapper[4715]: I1204 14:03:04.788727 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 14:03:04 crc kubenswrapper[4715]: I1204 14:03:04.791610 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 14:03:04 crc kubenswrapper[4715]: I1204 14:03:04.808971 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.146430 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.170118 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.170244 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.495661 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.766092 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.782974 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.900366 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 14:03:05 crc kubenswrapper[4715]: I1204 14:03:05.932413 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.041638 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.098165 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.218411 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.371768 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.385480 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.609886 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 14:03:06 crc kubenswrapper[4715]: I1204 14:03:06.723112 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.016937 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.208336 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.236223 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.498430 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.541772 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.552341 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.579871 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.597082 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.652761 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.654602 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.723013 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.739786 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.838268 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.857066 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.857217 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 14:03:07 crc kubenswrapper[4715]: I1204 14:03:07.980477 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.001663 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.487441 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.490522 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.537668 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.619602 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.633961 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.644899 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.669310 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.711497 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.792885 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.864895 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.902120 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 14:03:08 crc kubenswrapper[4715]: I1204 14:03:08.956622 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.053398 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.089642 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.201480 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.206868 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.215695 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.254164 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.319953 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.382344 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.417807 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.430408 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.465008 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.624298 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.626646 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.658651 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.683162 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.759419 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.837067 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.846419 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.930552 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 14:03:09 crc kubenswrapper[4715]: I1204 14:03:09.935315 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.104406 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.148676 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.183889 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.186744 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.206683 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.250594 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.336384 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.336901 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.352816 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.354909 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.393402 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.481446 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.573763 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.596804 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.647002 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.733351 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.739702 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.763164 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.763217 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.917282 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.929700 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.935924 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 14:03:10 crc kubenswrapper[4715]: I1204 14:03:10.957847 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.014239 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.034565 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.249753 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.256796 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.293487 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.394753 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.456915 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.555085 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.673276 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.784111 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.866436 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.955589 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 14:03:11 crc kubenswrapper[4715]: I1204 14:03:11.992631 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.058689 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.086128 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.095810 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.106992 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.134063 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.147280 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.167925 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.267731 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.467790 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.509290 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.676306 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.697163 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 14:03:12 crc kubenswrapper[4715]: I1204 14:03:12.888971 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.060383 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.128055 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.181135 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.206595 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.240714 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.258597 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.260099 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.303735 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.393416 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.401120 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.476700 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.492488 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.599510 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.618185 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.710175 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.835647 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 14:03:13 crc kubenswrapper[4715]: I1204 14:03:13.918073 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.033859 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.175626 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.183560 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.279993 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.333550 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.511663 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.521748 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.527442 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.615165 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.633004 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.654785 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.724664 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.840284 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.915462 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.927055 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.943543 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 14:03:14 crc kubenswrapper[4715]: I1204 14:03:14.984904 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.000542 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.005467 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-nrpl2"] Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.005529 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.005941 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.010660 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.042214 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.042198771 podStartE2EDuration="24.042198771s" podCreationTimestamp="2025-12-04 14:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:03:15.023499289 +0000 UTC m=+352.092217514" watchObservedRunningTime="2025-12-04 14:03:15.042198771 +0000 UTC m=+352.110916986" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.094181 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.116632 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.142273 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.173155 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.187206 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" path="/var/lib/kubelet/pods/3f118cab-ae82-49c9-90fb-e00688b70c5b/volumes" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.310837 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.333408 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.371708 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.373174 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.439752 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.455395 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.491300 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.498878 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.499120 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b3ae584922f448138f2eefee3b1f33bcd0dd1d8df86851152bbe70e4dfc42c4b" gracePeriod=5 Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.516347 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.552956 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.661516 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.770966 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.783467 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.805236 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 14:03:15 crc kubenswrapper[4715]: I1204 14:03:15.839740 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.001367 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.031492 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.040173 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.093193 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.207800 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.241211 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.243482 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.343641 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.412748 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.419788 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.629404 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.632016 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.759628 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.760702 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.768009 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.865060 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.865133 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.934968 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 14:03:16 crc kubenswrapper[4715]: I1204 14:03:16.958200 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064426 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9"] Dec 04 14:03:17 crc kubenswrapper[4715]: E1204 14:03:17.064678 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" containerName="oauth-openshift" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064692 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" containerName="oauth-openshift" Dec 04 14:03:17 crc kubenswrapper[4715]: E1204 14:03:17.064704 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064710 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 14:03:17 crc kubenswrapper[4715]: E1204 14:03:17.064730 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" containerName="installer" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064736 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" containerName="installer" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064836 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f118cab-ae82-49c9-90fb-e00688b70c5b" containerName="oauth-openshift" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064856 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c31b6e7-1c82-4ff4-822f-69c8aec95c43" containerName="installer" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.064866 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.065451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.069564 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.069771 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.069789 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.069914 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.071322 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.071525 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.071569 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.071597 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.071809 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.072243 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.073047 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.076391 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.081574 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9"] Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.084506 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.086512 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.091446 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.113429 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.149805 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.222224 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-audit-policies\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227618 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-router-certs\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227652 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-service-ca\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227675 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtgrh\" (UniqueName: \"kubernetes.io/projected/a2f6f989-7668-4a2f-99f0-db1f26f9b860-kube-api-access-xtgrh\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227702 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227747 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227767 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-login\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227790 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-error\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227820 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a2f6f989-7668-4a2f-99f0-db1f26f9b860-audit-dir\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.227894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-session\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.264453 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.328897 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.328962 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-login\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.328996 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-error\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329062 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329095 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a2f6f989-7668-4a2f-99f0-db1f26f9b860-audit-dir\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329119 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329176 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-session\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-audit-policies\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329226 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-router-certs\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329253 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-service-ca\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtgrh\" (UniqueName: \"kubernetes.io/projected/a2f6f989-7668-4a2f-99f0-db1f26f9b860-kube-api-access-xtgrh\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329317 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.329506 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a2f6f989-7668-4a2f-99f0-db1f26f9b860-audit-dir\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.330726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.331919 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.332236 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-service-ca\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.332270 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a2f6f989-7668-4a2f-99f0-db1f26f9b860-audit-policies\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.341299 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-error\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.341428 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.341480 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-session\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.341552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.341659 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-router-certs\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.342000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.342082 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.342528 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a2f6f989-7668-4a2f-99f0-db1f26f9b860-v4-0-config-user-template-login\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.349115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtgrh\" (UniqueName: \"kubernetes.io/projected/a2f6f989-7668-4a2f-99f0-db1f26f9b860-kube-api-access-xtgrh\") pod \"oauth-openshift-56f4fc5f47-pxjh9\" (UID: \"a2f6f989-7668-4a2f-99f0-db1f26f9b860\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.352923 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.382678 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.480198 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.485832 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.523835 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.540013 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.642023 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.667286 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.788939 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.794242 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4715]: I1204 14:03:17.805733 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9"] Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.135957 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.177175 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.192488 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.262670 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.354439 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.442622 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.447450 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.515715 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.547591 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.669773 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" event={"ID":"a2f6f989-7668-4a2f-99f0-db1f26f9b860","Type":"ContainerStarted","Data":"1f05786ed89ba431e6efcdab9c37efd55a6261b696d41dd02c5badf1cb654f01"} Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.669836 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" event={"ID":"a2f6f989-7668-4a2f-99f0-db1f26f9b860","Type":"ContainerStarted","Data":"9a04b91605dad19a66917fa142da5bcf6e8b6cbdfa3c4dd78bfdfa4641b47f2e"} Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.672972 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.676786 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.695277 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-56f4fc5f47-pxjh9" podStartSLOduration=70.695241596 podStartE2EDuration="1m10.695241596s" podCreationTimestamp="2025-12-04 14:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:03:18.690741815 +0000 UTC m=+355.759460030" watchObservedRunningTime="2025-12-04 14:03:18.695241596 +0000 UTC m=+355.763959831" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.807283 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.827910 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.874102 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.919260 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 14:03:18 crc kubenswrapper[4715]: I1204 14:03:18.985131 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.075736 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.139449 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.184079 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.250292 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.254232 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.310700 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.349942 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.446290 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.457890 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.501718 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.562840 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.771351 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.876606 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 14:03:19 crc kubenswrapper[4715]: I1204 14:03:19.972382 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 14:03:20 crc kubenswrapper[4715]: I1204 14:03:20.302344 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 14:03:20 crc kubenswrapper[4715]: I1204 14:03:20.666561 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 14:03:20 crc kubenswrapper[4715]: I1204 14:03:20.683613 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 14:03:20 crc kubenswrapper[4715]: I1204 14:03:20.683695 4715 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b3ae584922f448138f2eefee3b1f33bcd0dd1d8df86851152bbe70e4dfc42c4b" exitCode=137 Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.067018 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.067637 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184682 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184823 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184844 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184885 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184948 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.184988 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.185069 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.185096 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.185200 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.185213 4715 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.185223 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.185234 4715 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.196025 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.286094 4715 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.707617 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.707713 4715 scope.go:117] "RemoveContainer" containerID="b3ae584922f448138f2eefee3b1f33bcd0dd1d8df86851152bbe70e4dfc42c4b" Dec 04 14:03:21 crc kubenswrapper[4715]: I1204 14:03:21.707786 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 14:03:23 crc kubenswrapper[4715]: I1204 14:03:23.188075 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 14:03:34 crc kubenswrapper[4715]: I1204 14:03:34.971660 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 14:03:37 crc kubenswrapper[4715]: I1204 14:03:37.803568 4715 generic.go:334] "Generic (PLEG): container finished" podID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerID="12877f99a49b339195bab7253fd1c99bf310047d18f247a6025928c2ba3e8e98" exitCode=0 Dec 04 14:03:37 crc kubenswrapper[4715]: I1204 14:03:37.803615 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" event={"ID":"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c","Type":"ContainerDied","Data":"12877f99a49b339195bab7253fd1c99bf310047d18f247a6025928c2ba3e8e98"} Dec 04 14:03:37 crc kubenswrapper[4715]: I1204 14:03:37.804109 4715 scope.go:117] "RemoveContainer" containerID="12877f99a49b339195bab7253fd1c99bf310047d18f247a6025928c2ba3e8e98" Dec 04 14:03:38 crc kubenswrapper[4715]: I1204 14:03:38.757951 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:03:38 crc kubenswrapper[4715]: I1204 14:03:38.758269 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:03:38 crc kubenswrapper[4715]: I1204 14:03:38.808613 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" event={"ID":"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c","Type":"ContainerStarted","Data":"8078d9eda680cbee240f30a81201a21ca60aba78fcf86ceba459d9d06a9eeabf"} Dec 04 14:03:38 crc kubenswrapper[4715]: I1204 14:03:38.808911 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:03:38 crc kubenswrapper[4715]: I1204 14:03:38.810691 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:03:46 crc kubenswrapper[4715]: I1204 14:03:46.501247 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 14:03:56 crc kubenswrapper[4715]: I1204 14:03:56.636629 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.033533 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hb2z4"] Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.034086 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" podUID="261ac4be-400a-48b6-98de-ded0afc94441" containerName="controller-manager" containerID="cri-o://b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c" gracePeriod=30 Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.169118 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt"] Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.169631 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" podUID="ece56c58-71e4-421a-8dc0-04a543033972" containerName="route-controller-manager" containerID="cri-o://93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb" gracePeriod=30 Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.414026 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.518670 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.601540 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-proxy-ca-bundles\") pod \"261ac4be-400a-48b6-98de-ded0afc94441\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.601604 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-client-ca\") pod \"261ac4be-400a-48b6-98de-ded0afc94441\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.601663 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-config\") pod \"261ac4be-400a-48b6-98de-ded0afc94441\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.601743 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/261ac4be-400a-48b6-98de-ded0afc94441-serving-cert\") pod \"261ac4be-400a-48b6-98de-ded0afc94441\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.602633 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-config" (OuterVolumeSpecName: "config") pod "261ac4be-400a-48b6-98de-ded0afc94441" (UID: "261ac4be-400a-48b6-98de-ded0afc94441"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.602741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld2gq\" (UniqueName: \"kubernetes.io/projected/261ac4be-400a-48b6-98de-ded0afc94441-kube-api-access-ld2gq\") pod \"261ac4be-400a-48b6-98de-ded0afc94441\" (UID: \"261ac4be-400a-48b6-98de-ded0afc94441\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.602973 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdckm\" (UniqueName: \"kubernetes.io/projected/ece56c58-71e4-421a-8dc0-04a543033972-kube-api-access-rdckm\") pod \"ece56c58-71e4-421a-8dc0-04a543033972\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.603302 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-client-ca" (OuterVolumeSpecName: "client-ca") pod "261ac4be-400a-48b6-98de-ded0afc94441" (UID: "261ac4be-400a-48b6-98de-ded0afc94441"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.603465 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.603610 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.603908 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "261ac4be-400a-48b6-98de-ded0afc94441" (UID: "261ac4be-400a-48b6-98de-ded0afc94441"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.609546 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261ac4be-400a-48b6-98de-ded0afc94441-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "261ac4be-400a-48b6-98de-ded0afc94441" (UID: "261ac4be-400a-48b6-98de-ded0afc94441"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.609596 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece56c58-71e4-421a-8dc0-04a543033972-kube-api-access-rdckm" (OuterVolumeSpecName: "kube-api-access-rdckm") pod "ece56c58-71e4-421a-8dc0-04a543033972" (UID: "ece56c58-71e4-421a-8dc0-04a543033972"). InnerVolumeSpecName "kube-api-access-rdckm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.609685 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/261ac4be-400a-48b6-98de-ded0afc94441-kube-api-access-ld2gq" (OuterVolumeSpecName: "kube-api-access-ld2gq") pod "261ac4be-400a-48b6-98de-ded0afc94441" (UID: "261ac4be-400a-48b6-98de-ded0afc94441"). InnerVolumeSpecName "kube-api-access-ld2gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.704117 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-client-ca\") pod \"ece56c58-71e4-421a-8dc0-04a543033972\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.704565 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-config\") pod \"ece56c58-71e4-421a-8dc0-04a543033972\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.704657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece56c58-71e4-421a-8dc0-04a543033972-serving-cert\") pod \"ece56c58-71e4-421a-8dc0-04a543033972\" (UID: \"ece56c58-71e4-421a-8dc0-04a543033972\") " Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.704884 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/261ac4be-400a-48b6-98de-ded0afc94441-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.704985 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld2gq\" (UniqueName: \"kubernetes.io/projected/261ac4be-400a-48b6-98de-ded0afc94441-kube-api-access-ld2gq\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.705084 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdckm\" (UniqueName: \"kubernetes.io/projected/ece56c58-71e4-421a-8dc0-04a543033972-kube-api-access-rdckm\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.705192 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/261ac4be-400a-48b6-98de-ded0afc94441-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.705134 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-config" (OuterVolumeSpecName: "config") pod "ece56c58-71e4-421a-8dc0-04a543033972" (UID: "ece56c58-71e4-421a-8dc0-04a543033972"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.705147 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-client-ca" (OuterVolumeSpecName: "client-ca") pod "ece56c58-71e4-421a-8dc0-04a543033972" (UID: "ece56c58-71e4-421a-8dc0-04a543033972"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.707215 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece56c58-71e4-421a-8dc0-04a543033972-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ece56c58-71e4-421a-8dc0-04a543033972" (UID: "ece56c58-71e4-421a-8dc0-04a543033972"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.805928 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.805968 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece56c58-71e4-421a-8dc0-04a543033972-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.805978 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ece56c58-71e4-421a-8dc0-04a543033972-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.939625 4715 generic.go:334] "Generic (PLEG): container finished" podID="261ac4be-400a-48b6-98de-ded0afc94441" containerID="b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c" exitCode=0 Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.939689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" event={"ID":"261ac4be-400a-48b6-98de-ded0afc94441","Type":"ContainerDied","Data":"b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c"} Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.939711 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.939717 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hb2z4" event={"ID":"261ac4be-400a-48b6-98de-ded0afc94441","Type":"ContainerDied","Data":"8635a6fe214643cd0ed13b5df8d35ce1e2e5c8b9f61b0d1e9323fcd2f4cb9420"} Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.939728 4715 scope.go:117] "RemoveContainer" containerID="b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.948973 4715 generic.go:334] "Generic (PLEG): container finished" podID="ece56c58-71e4-421a-8dc0-04a543033972" containerID="93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb" exitCode=0 Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.949071 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" event={"ID":"ece56c58-71e4-421a-8dc0-04a543033972","Type":"ContainerDied","Data":"93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb"} Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.949619 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" event={"ID":"ece56c58-71e4-421a-8dc0-04a543033972","Type":"ContainerDied","Data":"9e8751e614b30ba3ebe75b217cb574036ee2cfb0148ab10cb3a56525045bba64"} Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.949107 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.963503 4715 scope.go:117] "RemoveContainer" containerID="b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c" Dec 04 14:04:03 crc kubenswrapper[4715]: E1204 14:04:03.964366 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c\": container with ID starting with b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c not found: ID does not exist" containerID="b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.964420 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c"} err="failed to get container status \"b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c\": rpc error: code = NotFound desc = could not find container \"b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c\": container with ID starting with b48843a7a9909c4898202b0eb0e49407ddca1d634265928b98f891c1cf6f515c not found: ID does not exist" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.964454 4715 scope.go:117] "RemoveContainer" containerID="93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.973124 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hb2z4"] Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.975674 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hb2z4"] Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.980267 4715 scope.go:117] "RemoveContainer" containerID="93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb" Dec 04 14:04:03 crc kubenswrapper[4715]: E1204 14:04:03.980793 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb\": container with ID starting with 93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb not found: ID does not exist" containerID="93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.980888 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb"} err="failed to get container status \"93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb\": rpc error: code = NotFound desc = could not find container \"93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb\": container with ID starting with 93c8cf0b7546523657cb2087a571e6a5219bd45fbdfeaf9aa59ff8932b743bbb not found: ID does not exist" Dec 04 14:04:03 crc kubenswrapper[4715]: I1204 14:04:03.999785 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt"] Dec 04 14:04:04 crc kubenswrapper[4715]: I1204 14:04:04.003288 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-drpkt"] Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.038082 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw"] Dec 04 14:04:05 crc kubenswrapper[4715]: E1204 14:04:05.038447 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261ac4be-400a-48b6-98de-ded0afc94441" containerName="controller-manager" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.038465 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="261ac4be-400a-48b6-98de-ded0afc94441" containerName="controller-manager" Dec 04 14:04:05 crc kubenswrapper[4715]: E1204 14:04:05.038484 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece56c58-71e4-421a-8dc0-04a543033972" containerName="route-controller-manager" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.038492 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece56c58-71e4-421a-8dc0-04a543033972" containerName="route-controller-manager" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.038609 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece56c58-71e4-421a-8dc0-04a543033972" containerName="route-controller-manager" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.038622 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="261ac4be-400a-48b6-98de-ded0afc94441" containerName="controller-manager" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.039132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.042143 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.043968 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-78c886458b-f8qnb"] Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.044865 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.045712 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.045814 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.045954 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.046051 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.046124 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.046406 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.048368 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.048640 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.048746 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.048807 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.049127 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.059873 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.061984 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-78c886458b-f8qnb"] Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.097073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw"] Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.186105 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="261ac4be-400a-48b6-98de-ded0afc94441" path="/var/lib/kubelet/pods/261ac4be-400a-48b6-98de-ded0afc94441/volumes" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.186702 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece56c58-71e4-421a-8dc0-04a543033972" path="/var/lib/kubelet/pods/ece56c58-71e4-421a-8dc0-04a543033972/volumes" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.222934 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-serving-cert\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.222986 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nvg5\" (UniqueName: \"kubernetes.io/projected/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-kube-api-access-8nvg5\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223017 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-proxy-ca-bundles\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223055 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-config\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223104 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-config\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223179 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-client-ca\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223223 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-client-ca\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223251 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7xd\" (UniqueName: \"kubernetes.io/projected/62e2405c-6b40-4538-b09f-574102af31b1-kube-api-access-wf7xd\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.223270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62e2405c-6b40-4538-b09f-574102af31b1-serving-cert\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.324342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nvg5\" (UniqueName: \"kubernetes.io/projected/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-kube-api-access-8nvg5\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.324875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-proxy-ca-bundles\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-config\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325166 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-config\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-client-ca\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325358 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-client-ca\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325493 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7xd\" (UniqueName: \"kubernetes.io/projected/62e2405c-6b40-4538-b09f-574102af31b1-kube-api-access-wf7xd\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325594 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62e2405c-6b40-4538-b09f-574102af31b1-serving-cert\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.325780 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-serving-cert\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.327697 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-client-ca\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.327717 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-client-ca\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.327713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-proxy-ca-bundles\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.328340 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-config\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.328602 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-config\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.332831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-serving-cert\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.333259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62e2405c-6b40-4538-b09f-574102af31b1-serving-cert\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.342751 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nvg5\" (UniqueName: \"kubernetes.io/projected/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-kube-api-access-8nvg5\") pod \"route-controller-manager-8697489c76-krlqw\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.345561 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7xd\" (UniqueName: \"kubernetes.io/projected/62e2405c-6b40-4538-b09f-574102af31b1-kube-api-access-wf7xd\") pod \"controller-manager-78c886458b-f8qnb\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.362322 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.368503 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.556800 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-78c886458b-f8qnb"] Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.599754 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw"] Dec 04 14:04:05 crc kubenswrapper[4715]: W1204 14:04:05.612547 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf7e24c8_39f2_45c4_b5e7_c615c4b6ccb4.slice/crio-44192908ac7278d8d78de47bd29e22ee61ad64e444e8db8901ec30b8de4b046e WatchSource:0}: Error finding container 44192908ac7278d8d78de47bd29e22ee61ad64e444e8db8901ec30b8de4b046e: Status 404 returned error can't find the container with id 44192908ac7278d8d78de47bd29e22ee61ad64e444e8db8901ec30b8de4b046e Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.970278 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" event={"ID":"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4","Type":"ContainerStarted","Data":"44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5"} Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.970616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" event={"ID":"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4","Type":"ContainerStarted","Data":"44192908ac7278d8d78de47bd29e22ee61ad64e444e8db8901ec30b8de4b046e"} Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.970931 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.973768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" event={"ID":"62e2405c-6b40-4538-b09f-574102af31b1","Type":"ContainerStarted","Data":"d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50"} Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.973812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" event={"ID":"62e2405c-6b40-4538-b09f-574102af31b1","Type":"ContainerStarted","Data":"d12c50b7e765192287e9eccc90c6ed298cdcec3f878e439c508a2a0abd0d6e37"} Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.974065 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.979377 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:05 crc kubenswrapper[4715]: I1204 14:04:05.992018 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" podStartSLOduration=2.992000122 podStartE2EDuration="2.992000122s" podCreationTimestamp="2025-12-04 14:04:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:04:05.989964594 +0000 UTC m=+403.058682809" watchObservedRunningTime="2025-12-04 14:04:05.992000122 +0000 UTC m=+403.060718337" Dec 04 14:04:06 crc kubenswrapper[4715]: I1204 14:04:06.011081 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" podStartSLOduration=3.011063066 podStartE2EDuration="3.011063066s" podCreationTimestamp="2025-12-04 14:04:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:04:06.010389616 +0000 UTC m=+403.079107831" watchObservedRunningTime="2025-12-04 14:04:06.011063066 +0000 UTC m=+403.079781301" Dec 04 14:04:06 crc kubenswrapper[4715]: I1204 14:04:06.222915 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:08 crc kubenswrapper[4715]: I1204 14:04:08.758294 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:04:08 crc kubenswrapper[4715]: I1204 14:04:08.758661 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.814611 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-h7f2b"] Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.816049 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.830200 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-h7f2b"] Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.962789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-registry-tls\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.963518 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.963650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.963802 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-registry-certificates\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.963945 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.964093 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-trusted-ca\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.964211 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-bound-sa-token\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.964362 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfk6m\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-kube-api-access-rfk6m\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:32 crc kubenswrapper[4715]: I1204 14:04:32.984896 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065363 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-registry-tls\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065432 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065520 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-registry-certificates\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065570 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-trusted-ca\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065591 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-bound-sa-token\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.065633 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfk6m\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-kube-api-access-rfk6m\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.066236 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.067463 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-trusted-ca\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.067498 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-registry-certificates\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.072346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-registry-tls\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.074713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.083452 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfk6m\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-kube-api-access-rfk6m\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.083714 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f-bound-sa-token\") pod \"image-registry-66df7c8f76-h7f2b\" (UID: \"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f\") " pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.132680 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:33 crc kubenswrapper[4715]: I1204 14:04:33.541829 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-h7f2b"] Dec 04 14:04:34 crc kubenswrapper[4715]: I1204 14:04:34.127206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" event={"ID":"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f","Type":"ContainerStarted","Data":"afb5dd583e14601ea00ea7d433bb9a3e7828ed8c4bdd587f98574b8e5f148ff4"} Dec 04 14:04:34 crc kubenswrapper[4715]: I1204 14:04:34.127590 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:34 crc kubenswrapper[4715]: I1204 14:04:34.127608 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" event={"ID":"afedeb65-4cb5-44e3-9ca2-f2ec7f2a575f","Type":"ContainerStarted","Data":"92dee07fa57367f21f0b9bfcd9b487cf2ab256caa23126f35dab887cc96cb455"} Dec 04 14:04:34 crc kubenswrapper[4715]: I1204 14:04:34.149827 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" podStartSLOduration=2.149808901 podStartE2EDuration="2.149808901s" podCreationTimestamp="2025-12-04 14:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:04:34.146907548 +0000 UTC m=+431.215625783" watchObservedRunningTime="2025-12-04 14:04:34.149808901 +0000 UTC m=+431.218527116" Dec 04 14:04:38 crc kubenswrapper[4715]: I1204 14:04:38.758875 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:04:38 crc kubenswrapper[4715]: I1204 14:04:38.759729 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:04:38 crc kubenswrapper[4715]: I1204 14:04:38.759960 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:04:38 crc kubenswrapper[4715]: I1204 14:04:38.761726 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a679b434089a9d4f4ce6522120aa8528b8358549b266643a6b526d3fdfd66568"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:04:38 crc kubenswrapper[4715]: I1204 14:04:38.762837 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://a679b434089a9d4f4ce6522120aa8528b8358549b266643a6b526d3fdfd66568" gracePeriod=600 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.163288 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="a679b434089a9d4f4ce6522120aa8528b8358549b266643a6b526d3fdfd66568" exitCode=0 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.163468 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"a679b434089a9d4f4ce6522120aa8528b8358549b266643a6b526d3fdfd66568"} Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.163711 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"6a6252ca8ea39b113b0e870cff0c9d7d47a3fe7fcf225413ba4908b3d3146389"} Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.163766 4715 scope.go:117] "RemoveContainer" containerID="a280228b315d8f9e7161b16cd3ed62ead9973b459cf5bd437d7b6512205e5042" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.592173 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9kjvl"] Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.592769 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9kjvl" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="registry-server" containerID="cri-o://c7d065177df835828ce208238792d767c634642b475ab21c8019215da0ad9c7a" gracePeriod=30 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.608601 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4tfcm"] Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.608859 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4tfcm" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="registry-server" containerID="cri-o://445542630ea202f186bd7cc4b877cdad1310c145bbbdac9eb61623c183ace444" gracePeriod=30 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.624265 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q9ctx"] Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.624492 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" containerID="cri-o://8078d9eda680cbee240f30a81201a21ca60aba78fcf86ceba459d9d06a9eeabf" gracePeriod=30 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.643515 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ssltf"] Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.644069 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ssltf" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="registry-server" containerID="cri-o://07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e" gracePeriod=30 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.653014 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7wkp2"] Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.653917 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.663256 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkbdv"] Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.663583 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vkbdv" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="registry-server" containerID="cri-o://0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6" gracePeriod=30 Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.669296 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7wkp2"] Dec 04 14:04:39 crc kubenswrapper[4715]: E1204 14:04:39.705187 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e is running failed: container process not found" containerID="07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:04:39 crc kubenswrapper[4715]: E1204 14:04:39.705627 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e is running failed: container process not found" containerID="07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:04:39 crc kubenswrapper[4715]: E1204 14:04:39.705854 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e is running failed: container process not found" containerID="07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:04:39 crc kubenswrapper[4715]: E1204 14:04:39.705884 4715 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-ssltf" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="registry-server" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.776721 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be7b5959-83d4-44ef-a596-f52b2e2a34f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.776781 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/be7b5959-83d4-44ef-a596-f52b2e2a34f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.776818 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk5dg\" (UniqueName: \"kubernetes.io/projected/be7b5959-83d4-44ef-a596-f52b2e2a34f7-kube-api-access-nk5dg\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.878584 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be7b5959-83d4-44ef-a596-f52b2e2a34f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.878918 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/be7b5959-83d4-44ef-a596-f52b2e2a34f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.879088 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk5dg\" (UniqueName: \"kubernetes.io/projected/be7b5959-83d4-44ef-a596-f52b2e2a34f7-kube-api-access-nk5dg\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.880925 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be7b5959-83d4-44ef-a596-f52b2e2a34f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.890835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/be7b5959-83d4-44ef-a596-f52b2e2a34f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.897949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk5dg\" (UniqueName: \"kubernetes.io/projected/be7b5959-83d4-44ef-a596-f52b2e2a34f7-kube-api-access-nk5dg\") pod \"marketplace-operator-79b997595-7wkp2\" (UID: \"be7b5959-83d4-44ef-a596-f52b2e2a34f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:39 crc kubenswrapper[4715]: I1204 14:04:39.985417 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.177466 4715 generic.go:334] "Generic (PLEG): container finished" podID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerID="8078d9eda680cbee240f30a81201a21ca60aba78fcf86ceba459d9d06a9eeabf" exitCode=0 Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.177685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" event={"ID":"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c","Type":"ContainerDied","Data":"8078d9eda680cbee240f30a81201a21ca60aba78fcf86ceba459d9d06a9eeabf"} Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.177836 4715 scope.go:117] "RemoveContainer" containerID="12877f99a49b339195bab7253fd1c99bf310047d18f247a6025928c2ba3e8e98" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.183877 4715 generic.go:334] "Generic (PLEG): container finished" podID="314da00a-3748-4149-b36c-8ced9d9f3018" containerID="c7d065177df835828ce208238792d767c634642b475ab21c8019215da0ad9c7a" exitCode=0 Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.183919 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kjvl" event={"ID":"314da00a-3748-4149-b36c-8ced9d9f3018","Type":"ContainerDied","Data":"c7d065177df835828ce208238792d767c634642b475ab21c8019215da0ad9c7a"} Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.190930 4715 generic.go:334] "Generic (PLEG): container finished" podID="b1fb6e69-52de-41b9-acae-df242cab542e" containerID="0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6" exitCode=0 Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.190990 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerDied","Data":"0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6"} Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.201459 4715 generic.go:334] "Generic (PLEG): container finished" podID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerID="07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e" exitCode=0 Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.201517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ssltf" event={"ID":"605514ff-63cc-4e34-a76e-44dd58be1c3e","Type":"ContainerDied","Data":"07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e"} Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.203560 4715 generic.go:334] "Generic (PLEG): container finished" podID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerID="445542630ea202f186bd7cc4b877cdad1310c145bbbdac9eb61623c183ace444" exitCode=0 Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.203594 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerDied","Data":"445542630ea202f186bd7cc4b877cdad1310c145bbbdac9eb61623c183ace444"} Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.413922 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7wkp2"] Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.547242 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q9ctx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.547332 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 14:04:40 crc kubenswrapper[4715]: E1204 14:04:40.768565 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6 is running failed: container process not found" containerID="0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:04:40 crc kubenswrapper[4715]: E1204 14:04:40.771542 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6 is running failed: container process not found" containerID="0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:04:40 crc kubenswrapper[4715]: E1204 14:04:40.771933 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6 is running failed: container process not found" containerID="0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:04:40 crc kubenswrapper[4715]: E1204 14:04:40.771997 4715 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-vkbdv" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="registry-server" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.808542 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.894783 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-utilities\") pod \"314da00a-3748-4149-b36c-8ced9d9f3018\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.894853 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z86nf\" (UniqueName: \"kubernetes.io/projected/314da00a-3748-4149-b36c-8ced9d9f3018-kube-api-access-z86nf\") pod \"314da00a-3748-4149-b36c-8ced9d9f3018\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.894940 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-catalog-content\") pod \"314da00a-3748-4149-b36c-8ced9d9f3018\" (UID: \"314da00a-3748-4149-b36c-8ced9d9f3018\") " Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.898231 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-utilities" (OuterVolumeSpecName: "utilities") pod "314da00a-3748-4149-b36c-8ced9d9f3018" (UID: "314da00a-3748-4149-b36c-8ced9d9f3018"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.912212 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314da00a-3748-4149-b36c-8ced9d9f3018-kube-api-access-z86nf" (OuterVolumeSpecName: "kube-api-access-z86nf") pod "314da00a-3748-4149-b36c-8ced9d9f3018" (UID: "314da00a-3748-4149-b36c-8ced9d9f3018"). InnerVolumeSpecName "kube-api-access-z86nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.965585 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "314da00a-3748-4149-b36c-8ced9d9f3018" (UID: "314da00a-3748-4149-b36c-8ced9d9f3018"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.996887 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z86nf\" (UniqueName: \"kubernetes.io/projected/314da00a-3748-4149-b36c-8ced9d9f3018-kube-api-access-z86nf\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.996916 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:40 crc kubenswrapper[4715]: I1204 14:04:40.996929 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da00a-3748-4149-b36c-8ced9d9f3018-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.080987 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.091347 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.094312 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.100611 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.198624 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-utilities\") pod \"605514ff-63cc-4e34-a76e-44dd58be1c3e\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.198888 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-operator-metrics\") pod \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199078 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-catalog-content\") pod \"605514ff-63cc-4e34-a76e-44dd58be1c3e\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199232 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzhxq\" (UniqueName: \"kubernetes.io/projected/e07d8a9f-8c02-4ac6-9e82-96564833487b-kube-api-access-zzhxq\") pod \"e07d8a9f-8c02-4ac6-9e82-96564833487b\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199334 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-trusted-ca\") pod \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199462 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-utilities" (OuterVolumeSpecName: "utilities") pod "605514ff-63cc-4e34-a76e-44dd58be1c3e" (UID: "605514ff-63cc-4e34-a76e-44dd58be1c3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199476 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-utilities\") pod \"b1fb6e69-52de-41b9-acae-df242cab542e\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199623 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dp8l\" (UniqueName: \"kubernetes.io/projected/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-kube-api-access-6dp8l\") pod \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\" (UID: \"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199663 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhbft\" (UniqueName: \"kubernetes.io/projected/605514ff-63cc-4e34-a76e-44dd58be1c3e-kube-api-access-hhbft\") pod \"605514ff-63cc-4e34-a76e-44dd58be1c3e\" (UID: \"605514ff-63cc-4e34-a76e-44dd58be1c3e\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199698 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-catalog-content\") pod \"b1fb6e69-52de-41b9-acae-df242cab542e\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199724 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78qvb\" (UniqueName: \"kubernetes.io/projected/b1fb6e69-52de-41b9-acae-df242cab542e-kube-api-access-78qvb\") pod \"b1fb6e69-52de-41b9-acae-df242cab542e\" (UID: \"b1fb6e69-52de-41b9-acae-df242cab542e\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.199764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-catalog-content\") pod \"e07d8a9f-8c02-4ac6-9e82-96564833487b\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.200671 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-utilities" (OuterVolumeSpecName: "utilities") pod "b1fb6e69-52de-41b9-acae-df242cab542e" (UID: "b1fb6e69-52de-41b9-acae-df242cab542e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.201157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" (UID: "0be0db16-0582-4dc7-8a68-f8a53b5b2a5c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.201113 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-utilities\") pod \"e07d8a9f-8c02-4ac6-9e82-96564833487b\" (UID: \"e07d8a9f-8c02-4ac6-9e82-96564833487b\") " Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.202125 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-utilities" (OuterVolumeSpecName: "utilities") pod "e07d8a9f-8c02-4ac6-9e82-96564833487b" (UID: "e07d8a9f-8c02-4ac6-9e82-96564833487b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.202365 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.202454 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.202541 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.202614 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.205319 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1fb6e69-52de-41b9-acae-df242cab542e-kube-api-access-78qvb" (OuterVolumeSpecName: "kube-api-access-78qvb") pod "b1fb6e69-52de-41b9-acae-df242cab542e" (UID: "b1fb6e69-52de-41b9-acae-df242cab542e"). InnerVolumeSpecName "kube-api-access-78qvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.205525 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e07d8a9f-8c02-4ac6-9e82-96564833487b-kube-api-access-zzhxq" (OuterVolumeSpecName: "kube-api-access-zzhxq") pod "e07d8a9f-8c02-4ac6-9e82-96564833487b" (UID: "e07d8a9f-8c02-4ac6-9e82-96564833487b"). InnerVolumeSpecName "kube-api-access-zzhxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.208299 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-kube-api-access-6dp8l" (OuterVolumeSpecName: "kube-api-access-6dp8l") pod "0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" (UID: "0be0db16-0582-4dc7-8a68-f8a53b5b2a5c"). InnerVolumeSpecName "kube-api-access-6dp8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.208364 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/605514ff-63cc-4e34-a76e-44dd58be1c3e-kube-api-access-hhbft" (OuterVolumeSpecName: "kube-api-access-hhbft") pod "605514ff-63cc-4e34-a76e-44dd58be1c3e" (UID: "605514ff-63cc-4e34-a76e-44dd58be1c3e"). InnerVolumeSpecName "kube-api-access-hhbft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.213271 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kjvl" event={"ID":"314da00a-3748-4149-b36c-8ced9d9f3018","Type":"ContainerDied","Data":"afb3da6ecb02f9164f8eaab66522d931738cf6bd6dfc68b30a33ce794bd6fe2b"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.213347 4715 scope.go:117] "RemoveContainer" containerID="c7d065177df835828ce208238792d767c634642b475ab21c8019215da0ad9c7a" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.213865 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kjvl" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.220718 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" (UID: "0be0db16-0582-4dc7-8a68-f8a53b5b2a5c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.226512 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "605514ff-63cc-4e34-a76e-44dd58be1c3e" (UID: "605514ff-63cc-4e34-a76e-44dd58be1c3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.234646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" event={"ID":"be7b5959-83d4-44ef-a596-f52b2e2a34f7","Type":"ContainerStarted","Data":"ef412d6dea5c931eecba13d93d43f5357438060c280d5552f13f40a3311ef934"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.234388 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7wkp2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" start-of-body= Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.235344 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" podUID="be7b5959-83d4-44ef-a596-f52b2e2a34f7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.235599 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" event={"ID":"be7b5959-83d4-44ef-a596-f52b2e2a34f7","Type":"ContainerStarted","Data":"4fd0eeda6636d0b94f9b5e5ca45d324d6cd1e71ff3662283878a3d70ec1fb449"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.235863 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.248378 4715 scope.go:117] "RemoveContainer" containerID="12bcc374284a033ae37f3498707f3dbdc59beac350eca70c3081a725d32f2f74" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.264092 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbdv" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.264436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbdv" event={"ID":"b1fb6e69-52de-41b9-acae-df242cab542e","Type":"ContainerDied","Data":"b97a535085ff9e6399af9ef198d1c8ff3fe49acdd103cb2dd661093e9481d92b"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.278594 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" podStartSLOduration=2.278567569 podStartE2EDuration="2.278567569s" podCreationTimestamp="2025-12-04 14:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:04:41.256989404 +0000 UTC m=+438.325707619" watchObservedRunningTime="2025-12-04 14:04:41.278567569 +0000 UTC m=+438.347285794" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.281802 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ssltf" event={"ID":"605514ff-63cc-4e34-a76e-44dd58be1c3e","Type":"ContainerDied","Data":"39cda0e41e6f49ff01a4d9e97457191b8533286596b866ec5e1099457601d755"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.281856 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9kjvl"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.281934 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ssltf" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.292822 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4tfcm" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.292893 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4tfcm" event={"ID":"e07d8a9f-8c02-4ac6-9e82-96564833487b","Type":"ContainerDied","Data":"fedec658ea178aa119927685b870d722f8b576fec17859a23dfde0bd5a8142c9"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.296961 4715 scope.go:117] "RemoveContainer" containerID="11a29881eecd867badb076befe7c43ed8449ce9ce8a5eedc129ce5c95e3c8de3" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.298576 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" event={"ID":"0be0db16-0582-4dc7-8a68-f8a53b5b2a5c","Type":"ContainerDied","Data":"d1549b17bcd037be46c25b264407b7d30773744d5d8977b3e56ec3a258a1051d"} Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.298651 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q9ctx" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.303741 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.303938 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/605514ff-63cc-4e34-a76e-44dd58be1c3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.304065 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzhxq\" (UniqueName: \"kubernetes.io/projected/e07d8a9f-8c02-4ac6-9e82-96564833487b-kube-api-access-zzhxq\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.304173 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dp8l\" (UniqueName: \"kubernetes.io/projected/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c-kube-api-access-6dp8l\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.304282 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhbft\" (UniqueName: \"kubernetes.io/projected/605514ff-63cc-4e34-a76e-44dd58be1c3e-kube-api-access-hhbft\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.304383 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78qvb\" (UniqueName: \"kubernetes.io/projected/b1fb6e69-52de-41b9-acae-df242cab542e-kube-api-access-78qvb\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.329424 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9kjvl"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.332258 4715 scope.go:117] "RemoveContainer" containerID="0b57f3d46d36686c854a22c6a689e84487e731408d9e09b5d06496854f9c08a6" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.338055 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ssltf"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.359388 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ssltf"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.362695 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q9ctx"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.362750 4715 scope.go:117] "RemoveContainer" containerID="f8a868bf64774ddd3a6a751558977f4f90476ae571a71aaf5de87982f974db4a" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.362947 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e07d8a9f-8c02-4ac6-9e82-96564833487b" (UID: "e07d8a9f-8c02-4ac6-9e82-96564833487b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.366129 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q9ctx"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.385204 4715 scope.go:117] "RemoveContainer" containerID="590d617354d567f122b71bdbb6e864319c22f447eaee003607db67a6c9286ec2" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.407638 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e07d8a9f-8c02-4ac6-9e82-96564833487b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.410473 4715 scope.go:117] "RemoveContainer" containerID="07b9ce1ef837a53f7b60e59a732ae45bb438039d8fbedc10695680702defcc5e" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.416488 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1fb6e69-52de-41b9-acae-df242cab542e" (UID: "b1fb6e69-52de-41b9-acae-df242cab542e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.430432 4715 scope.go:117] "RemoveContainer" containerID="b66947568c6b3a10392adc037f526779e42a312bd38ae9ac7d91772db16ca09f" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.459977 4715 scope.go:117] "RemoveContainer" containerID="c496fdaa4ee89b6db33e36b8dc3021a856e12f1bf6fb77232aba218cada6690d" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.484511 4715 scope.go:117] "RemoveContainer" containerID="445542630ea202f186bd7cc4b877cdad1310c145bbbdac9eb61623c183ace444" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.501821 4715 scope.go:117] "RemoveContainer" containerID="c92f3cc4c5cae56ef0d50a1306d6a54313a9bb4f333abbee440ebab60bd20bb5" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.508826 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1fb6e69-52de-41b9-acae-df242cab542e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.522955 4715 scope.go:117] "RemoveContainer" containerID="282d3af7215c4d3deccb315e2eae6b1574dced3b9163a508157591dad83a25b9" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.539174 4715 scope.go:117] "RemoveContainer" containerID="8078d9eda680cbee240f30a81201a21ca60aba78fcf86ceba459d9d06a9eeabf" Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.592555 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkbdv"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.596813 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vkbdv"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.617342 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4tfcm"] Dec 04 14:04:41 crc kubenswrapper[4715]: I1204 14:04:41.623908 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4tfcm"] Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.210719 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lsst9"] Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.210955 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.210976 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.210994 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211003 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211013 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211021 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211037 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211046 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211055 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211079 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211093 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211100 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211108 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211115 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211124 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211130 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211140 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211147 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211159 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211167 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211175 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211182 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211191 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211198 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="extract-utilities" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211209 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211216 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="extract-content" Dec 04 14:04:42 crc kubenswrapper[4715]: E1204 14:04:42.211227 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211234 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211338 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211354 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" containerName="marketplace-operator" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211362 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211373 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211382 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.211392 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" containerName="registry-server" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.213853 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.216399 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.227361 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lsst9"] Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.318261 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5md\" (UniqueName: \"kubernetes.io/projected/69d2a077-7d27-4c2f-82c5-d14628dbe954-kube-api-access-sd5md\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.318382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-catalog-content\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.318458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-utilities\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.319560 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7wkp2" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.419838 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5md\" (UniqueName: \"kubernetes.io/projected/69d2a077-7d27-4c2f-82c5-d14628dbe954-kube-api-access-sd5md\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.420275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-catalog-content\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.420419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-utilities\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.421118 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-catalog-content\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.421247 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-utilities\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.436939 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5md\" (UniqueName: \"kubernetes.io/projected/69d2a077-7d27-4c2f-82c5-d14628dbe954-kube-api-access-sd5md\") pod \"redhat-operators-lsst9\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.542652 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:42 crc kubenswrapper[4715]: I1204 14:04:42.805073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lsst9"] Dec 04 14:04:42 crc kubenswrapper[4715]: W1204 14:04:42.813267 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d2a077_7d27_4c2f_82c5_d14628dbe954.slice/crio-c2eb5e3b431281d045080e8e06976475e449ed7fc032d27e1fa6252e56ea177b WatchSource:0}: Error finding container c2eb5e3b431281d045080e8e06976475e449ed7fc032d27e1fa6252e56ea177b: Status 404 returned error can't find the container with id c2eb5e3b431281d045080e8e06976475e449ed7fc032d27e1fa6252e56ea177b Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.142404 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-78c886458b-f8qnb"] Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.142659 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" podUID="62e2405c-6b40-4538-b09f-574102af31b1" containerName="controller-manager" containerID="cri-o://d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50" gracePeriod=30 Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.188449 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be0db16-0582-4dc7-8a68-f8a53b5b2a5c" path="/var/lib/kubelet/pods/0be0db16-0582-4dc7-8a68-f8a53b5b2a5c/volumes" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.189015 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="314da00a-3748-4149-b36c-8ced9d9f3018" path="/var/lib/kubelet/pods/314da00a-3748-4149-b36c-8ced9d9f3018/volumes" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.189713 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="605514ff-63cc-4e34-a76e-44dd58be1c3e" path="/var/lib/kubelet/pods/605514ff-63cc-4e34-a76e-44dd58be1c3e/volumes" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.190900 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1fb6e69-52de-41b9-acae-df242cab542e" path="/var/lib/kubelet/pods/b1fb6e69-52de-41b9-acae-df242cab542e/volumes" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.191628 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e07d8a9f-8c02-4ac6-9e82-96564833487b" path="/var/lib/kubelet/pods/e07d8a9f-8c02-4ac6-9e82-96564833487b/volumes" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.205174 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2fvf6"] Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.208793 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: W1204 14:04:43.211220 4715 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 04 14:04:43 crc kubenswrapper[4715]: E1204 14:04:43.211275 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.221775 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2fvf6"] Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.318506 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw"] Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.324816 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" podUID="af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" containerName="route-controller-manager" containerID="cri-o://44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5" gracePeriod=30 Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.347566 4715 generic.go:334] "Generic (PLEG): container finished" podID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerID="94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1" exitCode=0 Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.348494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lsst9" event={"ID":"69d2a077-7d27-4c2f-82c5-d14628dbe954","Type":"ContainerDied","Data":"94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1"} Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.348575 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lsst9" event={"ID":"69d2a077-7d27-4c2f-82c5-d14628dbe954","Type":"ContainerStarted","Data":"c2eb5e3b431281d045080e8e06976475e449ed7fc032d27e1fa6252e56ea177b"} Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.360635 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-utilities\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.361527 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjwrp\" (UniqueName: \"kubernetes.io/projected/4968998b-fada-4949-beb1-982de165dc3c-kube-api-access-xjwrp\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.361788 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-catalog-content\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.463786 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-utilities\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.465290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-utilities\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.465453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjwrp\" (UniqueName: \"kubernetes.io/projected/4968998b-fada-4949-beb1-982de165dc3c-kube-api-access-xjwrp\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.465905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-catalog-content\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.466397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-catalog-content\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.491550 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjwrp\" (UniqueName: \"kubernetes.io/projected/4968998b-fada-4949-beb1-982de165dc3c-kube-api-access-xjwrp\") pod \"certified-operators-2fvf6\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.776761 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.871891 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-client-ca\") pod \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.871947 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-serving-cert\") pod \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.872038 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-config\") pod \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.872115 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nvg5\" (UniqueName: \"kubernetes.io/projected/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-kube-api-access-8nvg5\") pod \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\" (UID: \"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4\") " Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.872947 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-client-ca" (OuterVolumeSpecName: "client-ca") pod "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" (UID: "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.873595 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-config" (OuterVolumeSpecName: "config") pod "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" (UID: "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.878020 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-kube-api-access-8nvg5" (OuterVolumeSpecName: "kube-api-access-8nvg5") pod "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" (UID: "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4"). InnerVolumeSpecName "kube-api-access-8nvg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.878684 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" (UID: "af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.973682 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nvg5\" (UniqueName: \"kubernetes.io/projected/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-kube-api-access-8nvg5\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.973715 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.973726 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:43 crc kubenswrapper[4715]: I1204 14:04:43.973736 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.226011 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.356178 4715 generic.go:334] "Generic (PLEG): container finished" podID="af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" containerID="44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5" exitCode=0 Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.356266 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.356310 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" event={"ID":"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4","Type":"ContainerDied","Data":"44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5"} Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.356391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw" event={"ID":"af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4","Type":"ContainerDied","Data":"44192908ac7278d8d78de47bd29e22ee61ad64e444e8db8901ec30b8de4b046e"} Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.356423 4715 scope.go:117] "RemoveContainer" containerID="44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.360836 4715 generic.go:334] "Generic (PLEG): container finished" podID="62e2405c-6b40-4538-b09f-574102af31b1" containerID="d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50" exitCode=0 Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.360863 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.360893 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" event={"ID":"62e2405c-6b40-4538-b09f-574102af31b1","Type":"ContainerDied","Data":"d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50"} Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.361844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78c886458b-f8qnb" event={"ID":"62e2405c-6b40-4538-b09f-574102af31b1","Type":"ContainerDied","Data":"d12c50b7e765192287e9eccc90c6ed298cdcec3f878e439c508a2a0abd0d6e37"} Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.384856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-client-ca\") pod \"62e2405c-6b40-4538-b09f-574102af31b1\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.384925 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf7xd\" (UniqueName: \"kubernetes.io/projected/62e2405c-6b40-4538-b09f-574102af31b1-kube-api-access-wf7xd\") pod \"62e2405c-6b40-4538-b09f-574102af31b1\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.384975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-config\") pod \"62e2405c-6b40-4538-b09f-574102af31b1\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.384994 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-proxy-ca-bundles\") pod \"62e2405c-6b40-4538-b09f-574102af31b1\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.385090 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62e2405c-6b40-4538-b09f-574102af31b1-serving-cert\") pod \"62e2405c-6b40-4538-b09f-574102af31b1\" (UID: \"62e2405c-6b40-4538-b09f-574102af31b1\") " Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.385842 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "62e2405c-6b40-4538-b09f-574102af31b1" (UID: "62e2405c-6b40-4538-b09f-574102af31b1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.385852 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-client-ca" (OuterVolumeSpecName: "client-ca") pod "62e2405c-6b40-4538-b09f-574102af31b1" (UID: "62e2405c-6b40-4538-b09f-574102af31b1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.386243 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-config" (OuterVolumeSpecName: "config") pod "62e2405c-6b40-4538-b09f-574102af31b1" (UID: "62e2405c-6b40-4538-b09f-574102af31b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.394293 4715 scope.go:117] "RemoveContainer" containerID="44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5" Dec 04 14:04:44 crc kubenswrapper[4715]: E1204 14:04:44.394973 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5\": container with ID starting with 44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5 not found: ID does not exist" containerID="44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.395003 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5"} err="failed to get container status \"44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5\": rpc error: code = NotFound desc = could not find container \"44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5\": container with ID starting with 44980e3c8863bdc46d7c6627b690919923bf5894a753618248267bf93a4022f5 not found: ID does not exist" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.395028 4715 scope.go:117] "RemoveContainer" containerID="d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.399369 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.401013 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw"] Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.402463 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.404733 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8697489c76-krlqw"] Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.417574 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e2405c-6b40-4538-b09f-574102af31b1-kube-api-access-wf7xd" (OuterVolumeSpecName: "kube-api-access-wf7xd") pod "62e2405c-6b40-4538-b09f-574102af31b1" (UID: "62e2405c-6b40-4538-b09f-574102af31b1"). InnerVolumeSpecName "kube-api-access-wf7xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.427755 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e2405c-6b40-4538-b09f-574102af31b1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "62e2405c-6b40-4538-b09f-574102af31b1" (UID: "62e2405c-6b40-4538-b09f-574102af31b1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.486223 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.486257 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf7xd\" (UniqueName: \"kubernetes.io/projected/62e2405c-6b40-4538-b09f-574102af31b1-kube-api-access-wf7xd\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.486269 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.486279 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62e2405c-6b40-4538-b09f-574102af31b1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.486288 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62e2405c-6b40-4538-b09f-574102af31b1-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.501025 4715 scope.go:117] "RemoveContainer" containerID="d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50" Dec 04 14:04:44 crc kubenswrapper[4715]: E1204 14:04:44.501654 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50\": container with ID starting with d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50 not found: ID does not exist" containerID="d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.501715 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50"} err="failed to get container status \"d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50\": rpc error: code = NotFound desc = could not find container \"d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50\": container with ID starting with d3aaf942089304952fcb0f772f421466c756ca03ef7c08792efee5b503c2dd50 not found: ID does not exist" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.621502 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h5dgv"] Dec 04 14:04:44 crc kubenswrapper[4715]: E1204 14:04:44.621786 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e2405c-6b40-4538-b09f-574102af31b1" containerName="controller-manager" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.621804 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e2405c-6b40-4538-b09f-574102af31b1" containerName="controller-manager" Dec 04 14:04:44 crc kubenswrapper[4715]: E1204 14:04:44.621819 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" containerName="route-controller-manager" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.621826 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" containerName="route-controller-manager" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.621931 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" containerName="route-controller-manager" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.621941 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e2405c-6b40-4538-b09f-574102af31b1" containerName="controller-manager" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.622741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.626584 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.627738 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h5dgv"] Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.724286 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-78c886458b-f8qnb"] Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.731540 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-78c886458b-f8qnb"] Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.746141 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2fvf6"] Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.790882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-catalog-content\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.790976 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phwlm\" (UniqueName: \"kubernetes.io/projected/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-kube-api-access-phwlm\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.791015 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-utilities\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.892623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-catalog-content\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.892718 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phwlm\" (UniqueName: \"kubernetes.io/projected/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-kube-api-access-phwlm\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.892755 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-utilities\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.893450 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-catalog-content\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.893651 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-utilities\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.922043 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phwlm\" (UniqueName: \"kubernetes.io/projected/26ec7fe9-b52f-4211-a5bf-ab16f43cb7af-kube-api-access-phwlm\") pod \"community-operators-h5dgv\" (UID: \"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af\") " pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:44 crc kubenswrapper[4715]: I1204 14:04:44.950685 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.074930 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl"] Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.075949 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.080167 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.080497 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.080666 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.080854 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.080982 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.081158 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.104568 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85bf86677c-j2tpn"] Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.108530 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.108949 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl"] Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.112642 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.114147 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.114804 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.115199 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85bf86677c-j2tpn"] Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.115675 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.115817 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.127289 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.150574 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.189962 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e2405c-6b40-4538-b09f-574102af31b1" path="/var/lib/kubelet/pods/62e2405c-6b40-4538-b09f-574102af31b1/volumes" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.191857 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4" path="/var/lib/kubelet/pods/af7e24c8-39f2-45c4-b5e7-c615c4b6ccb4/volumes" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.196402 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dswqj\" (UniqueName: \"kubernetes.io/projected/bef8dc0d-2bfd-472d-9197-61b0b34f5189-kube-api-access-dswqj\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.196445 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef8dc0d-2bfd-472d-9197-61b0b34f5189-config\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.196534 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef8dc0d-2bfd-472d-9197-61b0b34f5189-serving-cert\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.196614 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef8dc0d-2bfd-472d-9197-61b0b34f5189-client-ca\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.236311 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h5dgv"] Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.298486 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef8dc0d-2bfd-472d-9197-61b0b34f5189-serving-cert\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.298905 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwrt6\" (UniqueName: \"kubernetes.io/projected/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-kube-api-access-qwrt6\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.298998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef8dc0d-2bfd-472d-9197-61b0b34f5189-client-ca\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.299014 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-config\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.299123 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dswqj\" (UniqueName: \"kubernetes.io/projected/bef8dc0d-2bfd-472d-9197-61b0b34f5189-kube-api-access-dswqj\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.299164 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef8dc0d-2bfd-472d-9197-61b0b34f5189-config\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.299191 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-client-ca\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.299211 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-serving-cert\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.299230 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-proxy-ca-bundles\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.301318 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef8dc0d-2bfd-472d-9197-61b0b34f5189-client-ca\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.301841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef8dc0d-2bfd-472d-9197-61b0b34f5189-config\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.312069 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef8dc0d-2bfd-472d-9197-61b0b34f5189-serving-cert\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.319242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dswqj\" (UniqueName: \"kubernetes.io/projected/bef8dc0d-2bfd-472d-9197-61b0b34f5189-kube-api-access-dswqj\") pod \"route-controller-manager-75f88c5b7c-xwvsl\" (UID: \"bef8dc0d-2bfd-472d-9197-61b0b34f5189\") " pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.383018 4715 generic.go:334] "Generic (PLEG): container finished" podID="4968998b-fada-4949-beb1-982de165dc3c" containerID="e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d" exitCode=0 Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.383221 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2fvf6" event={"ID":"4968998b-fada-4949-beb1-982de165dc3c","Type":"ContainerDied","Data":"e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d"} Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.383761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2fvf6" event={"ID":"4968998b-fada-4949-beb1-982de165dc3c","Type":"ContainerStarted","Data":"9ac5ed4c66ecd8e8a00e9a52ce5c688e7eef85de547c12443e25a755f512206f"} Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.387967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5dgv" event={"ID":"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af","Type":"ContainerStarted","Data":"8f69c97ef3c31798e273346f30be01174a7f7869660d416b0c95eca208c0646b"} Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.398131 4715 generic.go:334] "Generic (PLEG): container finished" podID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerID="d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b" exitCode=0 Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.398254 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lsst9" event={"ID":"69d2a077-7d27-4c2f-82c5-d14628dbe954","Type":"ContainerDied","Data":"d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b"} Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.400082 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwrt6\" (UniqueName: \"kubernetes.io/projected/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-kube-api-access-qwrt6\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.400161 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-config\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.400230 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-client-ca\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.400259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-serving-cert\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.400287 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-proxy-ca-bundles\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.401405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-client-ca\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.401626 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-proxy-ca-bundles\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.405874 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-serving-cert\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.406171 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-config\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.414308 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.421662 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwrt6\" (UniqueName: \"kubernetes.io/projected/159b0486-8e3d-4d05-a42f-fcf4fd36f6f1-kube-api-access-qwrt6\") pod \"controller-manager-85bf86677c-j2tpn\" (UID: \"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1\") " pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.437472 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.861838 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl"] Dec 04 14:04:45 crc kubenswrapper[4715]: I1204 14:04:45.958856 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85bf86677c-j2tpn"] Dec 04 14:04:45 crc kubenswrapper[4715]: W1204 14:04:45.971834 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod159b0486_8e3d_4d05_a42f_fcf4fd36f6f1.slice/crio-385677da7a69433aa2fb675c6aae8f1bb3a55ab7171f48fa1c012e5c35bbd86a WatchSource:0}: Error finding container 385677da7a69433aa2fb675c6aae8f1bb3a55ab7171f48fa1c012e5c35bbd86a: Status 404 returned error can't find the container with id 385677da7a69433aa2fb675c6aae8f1bb3a55ab7171f48fa1c012e5c35bbd86a Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.010701 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4lg6p"] Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.011930 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.019110 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.026275 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lg6p"] Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.109836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d093b14-52c6-442a-901d-895fa5b9b721-catalog-content\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.110418 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b9wt\" (UniqueName: \"kubernetes.io/projected/0d093b14-52c6-442a-901d-895fa5b9b721-kube-api-access-5b9wt\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.110456 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d093b14-52c6-442a-901d-895fa5b9b721-utilities\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.212771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d093b14-52c6-442a-901d-895fa5b9b721-catalog-content\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.212870 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b9wt\" (UniqueName: \"kubernetes.io/projected/0d093b14-52c6-442a-901d-895fa5b9b721-kube-api-access-5b9wt\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.212901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d093b14-52c6-442a-901d-895fa5b9b721-utilities\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.213881 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d093b14-52c6-442a-901d-895fa5b9b721-catalog-content\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.214244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d093b14-52c6-442a-901d-895fa5b9b721-utilities\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.239585 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b9wt\" (UniqueName: \"kubernetes.io/projected/0d093b14-52c6-442a-901d-895fa5b9b721-kube-api-access-5b9wt\") pod \"redhat-marketplace-4lg6p\" (UID: \"0d093b14-52c6-442a-901d-895fa5b9b721\") " pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.338908 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.413318 4715 generic.go:334] "Generic (PLEG): container finished" podID="26ec7fe9-b52f-4211-a5bf-ab16f43cb7af" containerID="23d08ac8700317b51c52dc9525a21db2e3d4c493fb839398e24bd3a32c5593f1" exitCode=0 Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.413811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5dgv" event={"ID":"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af","Type":"ContainerDied","Data":"23d08ac8700317b51c52dc9525a21db2e3d4c493fb839398e24bd3a32c5593f1"} Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.427843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" event={"ID":"bef8dc0d-2bfd-472d-9197-61b0b34f5189","Type":"ContainerStarted","Data":"77132b526b85ed4c3de0afcfb28b7bfa7c84cf57cfe8de8b2b1afc7f9ba8300e"} Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.428462 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" event={"ID":"bef8dc0d-2bfd-472d-9197-61b0b34f5189","Type":"ContainerStarted","Data":"90844bce8f161856d763402d15cb37f10598f9adbdb9ebd1a74de8f13a23b58c"} Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.428501 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.434771 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.438413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" event={"ID":"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1","Type":"ContainerStarted","Data":"a8181ac17c05028a757431043e66d1b038d70359503e9a08dac1d9723e5cf8ec"} Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.438489 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" event={"ID":"159b0486-8e3d-4d05-a42f-fcf4fd36f6f1","Type":"ContainerStarted","Data":"385677da7a69433aa2fb675c6aae8f1bb3a55ab7171f48fa1c012e5c35bbd86a"} Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.439226 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.452841 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.469130 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75f88c5b7c-xwvsl" podStartSLOduration=3.469107961 podStartE2EDuration="3.469107961s" podCreationTimestamp="2025-12-04 14:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:04:46.465528129 +0000 UTC m=+443.534246364" watchObservedRunningTime="2025-12-04 14:04:46.469107961 +0000 UTC m=+443.537826186" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.493436 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85bf86677c-j2tpn" podStartSLOduration=3.493417654 podStartE2EDuration="3.493417654s" podCreationTimestamp="2025-12-04 14:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:04:46.492481287 +0000 UTC m=+443.561199512" watchObservedRunningTime="2025-12-04 14:04:46.493417654 +0000 UTC m=+443.562135869" Dec 04 14:04:46 crc kubenswrapper[4715]: I1204 14:04:46.680736 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lg6p"] Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.447052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lsst9" event={"ID":"69d2a077-7d27-4c2f-82c5-d14628dbe954","Type":"ContainerStarted","Data":"ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102"} Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.448762 4715 generic.go:334] "Generic (PLEG): container finished" podID="0d093b14-52c6-442a-901d-895fa5b9b721" containerID="0389b929cc79a1c4360d821e414e34f1b1bb031cdbcd35c09bba1244d8a5569e" exitCode=0 Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.448850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lg6p" event={"ID":"0d093b14-52c6-442a-901d-895fa5b9b721","Type":"ContainerDied","Data":"0389b929cc79a1c4360d821e414e34f1b1bb031cdbcd35c09bba1244d8a5569e"} Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.448896 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lg6p" event={"ID":"0d093b14-52c6-442a-901d-895fa5b9b721","Type":"ContainerStarted","Data":"05197df9ee3e029a3bb64a7f57dd4f4b4a67edb0ef770c476df265eb7f87c951"} Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.451170 4715 generic.go:334] "Generic (PLEG): container finished" podID="4968998b-fada-4949-beb1-982de165dc3c" containerID="bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485" exitCode=0 Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.451233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2fvf6" event={"ID":"4968998b-fada-4949-beb1-982de165dc3c","Type":"ContainerDied","Data":"bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485"} Dec 04 14:04:47 crc kubenswrapper[4715]: I1204 14:04:47.469568 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lsst9" podStartSLOduration=2.70500233 podStartE2EDuration="5.469552077s" podCreationTimestamp="2025-12-04 14:04:42 +0000 UTC" firstStartedPulling="2025-12-04 14:04:43.35023661 +0000 UTC m=+440.418954825" lastFinishedPulling="2025-12-04 14:04:46.114786357 +0000 UTC m=+443.183504572" observedRunningTime="2025-12-04 14:04:47.468822447 +0000 UTC m=+444.537540662" watchObservedRunningTime="2025-12-04 14:04:47.469552077 +0000 UTC m=+444.538270292" Dec 04 14:04:48 crc kubenswrapper[4715]: I1204 14:04:48.464418 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2fvf6" event={"ID":"4968998b-fada-4949-beb1-982de165dc3c","Type":"ContainerStarted","Data":"34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18"} Dec 04 14:04:48 crc kubenswrapper[4715]: I1204 14:04:48.468294 4715 generic.go:334] "Generic (PLEG): container finished" podID="26ec7fe9-b52f-4211-a5bf-ab16f43cb7af" containerID="ad4dde7cc0c8a97efe231520706d8ccc41a64004d53c34df105bc6eb15ad969b" exitCode=0 Dec 04 14:04:48 crc kubenswrapper[4715]: I1204 14:04:48.468387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5dgv" event={"ID":"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af","Type":"ContainerDied","Data":"ad4dde7cc0c8a97efe231520706d8ccc41a64004d53c34df105bc6eb15ad969b"} Dec 04 14:04:48 crc kubenswrapper[4715]: I1204 14:04:48.488097 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2fvf6" podStartSLOduration=2.77954846 podStartE2EDuration="5.48807716s" podCreationTimestamp="2025-12-04 14:04:43 +0000 UTC" firstStartedPulling="2025-12-04 14:04:45.386171493 +0000 UTC m=+442.454889708" lastFinishedPulling="2025-12-04 14:04:48.094700193 +0000 UTC m=+445.163418408" observedRunningTime="2025-12-04 14:04:48.483013895 +0000 UTC m=+445.551732110" watchObservedRunningTime="2025-12-04 14:04:48.48807716 +0000 UTC m=+445.556795375" Dec 04 14:04:49 crc kubenswrapper[4715]: I1204 14:04:49.478736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5dgv" event={"ID":"26ec7fe9-b52f-4211-a5bf-ab16f43cb7af","Type":"ContainerStarted","Data":"c548e74a6bfa412b9d8664c7847afaf93d5f6fba15615960836096fec442b0b7"} Dec 04 14:04:49 crc kubenswrapper[4715]: I1204 14:04:49.480812 4715 generic.go:334] "Generic (PLEG): container finished" podID="0d093b14-52c6-442a-901d-895fa5b9b721" containerID="fd00ce6c847ff6d4608cd49314e0ed627385784fe4e1e1c5ba32d917c24ad1a9" exitCode=0 Dec 04 14:04:49 crc kubenswrapper[4715]: I1204 14:04:49.480877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lg6p" event={"ID":"0d093b14-52c6-442a-901d-895fa5b9b721","Type":"ContainerDied","Data":"fd00ce6c847ff6d4608cd49314e0ed627385784fe4e1e1c5ba32d917c24ad1a9"} Dec 04 14:04:49 crc kubenswrapper[4715]: I1204 14:04:49.500309 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h5dgv" podStartSLOduration=2.964236077 podStartE2EDuration="5.500291701s" podCreationTimestamp="2025-12-04 14:04:44 +0000 UTC" firstStartedPulling="2025-12-04 14:04:46.42595977 +0000 UTC m=+443.494677995" lastFinishedPulling="2025-12-04 14:04:48.962015404 +0000 UTC m=+446.030733619" observedRunningTime="2025-12-04 14:04:49.495464223 +0000 UTC m=+446.564182448" watchObservedRunningTime="2025-12-04 14:04:49.500291701 +0000 UTC m=+446.569009916" Dec 04 14:04:51 crc kubenswrapper[4715]: I1204 14:04:51.498879 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lg6p" event={"ID":"0d093b14-52c6-442a-901d-895fa5b9b721","Type":"ContainerStarted","Data":"15f7ea08dfa9d5bc1cbee08d3dfb09ffea84aeca7556434e15486be0ff99240f"} Dec 04 14:04:51 crc kubenswrapper[4715]: I1204 14:04:51.527620 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4lg6p" podStartSLOduration=3.452064483 podStartE2EDuration="6.527594778s" podCreationTimestamp="2025-12-04 14:04:45 +0000 UTC" firstStartedPulling="2025-12-04 14:04:47.459909482 +0000 UTC m=+444.528627697" lastFinishedPulling="2025-12-04 14:04:50.535439777 +0000 UTC m=+447.604157992" observedRunningTime="2025-12-04 14:04:51.523276825 +0000 UTC m=+448.591995060" watchObservedRunningTime="2025-12-04 14:04:51.527594778 +0000 UTC m=+448.596313003" Dec 04 14:04:52 crc kubenswrapper[4715]: I1204 14:04:52.543516 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:52 crc kubenswrapper[4715]: I1204 14:04:52.543910 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:04:53 crc kubenswrapper[4715]: I1204 14:04:53.140912 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-h7f2b" Dec 04 14:04:53 crc kubenswrapper[4715]: I1204 14:04:53.199839 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-79f5b"] Dec 04 14:04:53 crc kubenswrapper[4715]: I1204 14:04:53.590130 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lsst9" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="registry-server" probeResult="failure" output=< Dec 04 14:04:53 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 04 14:04:53 crc kubenswrapper[4715]: > Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.403151 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.403536 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.453026 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.564660 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.951539 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.952494 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:54 crc kubenswrapper[4715]: I1204 14:04:54.997108 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:55 crc kubenswrapper[4715]: I1204 14:04:55.567186 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h5dgv" Dec 04 14:04:56 crc kubenswrapper[4715]: I1204 14:04:56.339863 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:56 crc kubenswrapper[4715]: I1204 14:04:56.342724 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:56 crc kubenswrapper[4715]: I1204 14:04:56.395254 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:04:56 crc kubenswrapper[4715]: I1204 14:04:56.571809 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4lg6p" Dec 04 14:05:02 crc kubenswrapper[4715]: I1204 14:05:02.588517 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:05:02 crc kubenswrapper[4715]: I1204 14:05:02.641276 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:05:18 crc kubenswrapper[4715]: I1204 14:05:18.250610 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" podUID="678577cf-42b5-4f4a-84d2-9be8e1d7b753" containerName="registry" containerID="cri-o://49b6854642573b40395b960ad4f35c676bcb6ed7ed8dfe8ddfde8e1f1027012d" gracePeriod=30 Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.086726 4715 generic.go:334] "Generic (PLEG): container finished" podID="678577cf-42b5-4f4a-84d2-9be8e1d7b753" containerID="49b6854642573b40395b960ad4f35c676bcb6ed7ed8dfe8ddfde8e1f1027012d" exitCode=0 Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.086985 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" event={"ID":"678577cf-42b5-4f4a-84d2-9be8e1d7b753","Type":"ContainerDied","Data":"49b6854642573b40395b960ad4f35c676bcb6ed7ed8dfe8ddfde8e1f1027012d"} Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.268935 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.372673 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsn5x\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-kube-api-access-dsn5x\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.372771 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-trusted-ca\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.372830 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/678577cf-42b5-4f4a-84d2-9be8e1d7b753-ca-trust-extracted\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.372995 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/678577cf-42b5-4f4a-84d2-9be8e1d7b753-installation-pull-secrets\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.373052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-bound-sa-token\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.373093 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-certificates\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.373261 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.373310 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-tls\") pod \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\" (UID: \"678577cf-42b5-4f4a-84d2-9be8e1d7b753\") " Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.373653 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.373827 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.374762 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.381521 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.381688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678577cf-42b5-4f4a-84d2-9be8e1d7b753-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.381999 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.384080 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.387281 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-kube-api-access-dsn5x" (OuterVolumeSpecName: "kube-api-access-dsn5x") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "kube-api-access-dsn5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.392399 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/678577cf-42b5-4f4a-84d2-9be8e1d7b753-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "678577cf-42b5-4f4a-84d2-9be8e1d7b753" (UID: "678577cf-42b5-4f4a-84d2-9be8e1d7b753"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.474983 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/678577cf-42b5-4f4a-84d2-9be8e1d7b753-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.475022 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/678577cf-42b5-4f4a-84d2-9be8e1d7b753-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.475053 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.475066 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.475078 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:19 crc kubenswrapper[4715]: I1204 14:05:19.475091 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsn5x\" (UniqueName: \"kubernetes.io/projected/678577cf-42b5-4f4a-84d2-9be8e1d7b753-kube-api-access-dsn5x\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:20 crc kubenswrapper[4715]: I1204 14:05:20.094295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" event={"ID":"678577cf-42b5-4f4a-84d2-9be8e1d7b753","Type":"ContainerDied","Data":"d4a4f725cbfa51133d6af3f4682fb48638e98ca9ae915af259a2ee93cd7cc8d1"} Dec 04 14:05:20 crc kubenswrapper[4715]: I1204 14:05:20.094378 4715 scope.go:117] "RemoveContainer" containerID="49b6854642573b40395b960ad4f35c676bcb6ed7ed8dfe8ddfde8e1f1027012d" Dec 04 14:05:20 crc kubenswrapper[4715]: I1204 14:05:20.094740 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-79f5b" Dec 04 14:05:20 crc kubenswrapper[4715]: I1204 14:05:20.151239 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-79f5b"] Dec 04 14:05:20 crc kubenswrapper[4715]: I1204 14:05:20.155935 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-79f5b"] Dec 04 14:05:21 crc kubenswrapper[4715]: I1204 14:05:21.191739 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="678577cf-42b5-4f4a-84d2-9be8e1d7b753" path="/var/lib/kubelet/pods/678577cf-42b5-4f4a-84d2-9be8e1d7b753/volumes" Dec 04 14:07:08 crc kubenswrapper[4715]: I1204 14:07:08.758214 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:07:08 crc kubenswrapper[4715]: I1204 14:07:08.758725 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:07:38 crc kubenswrapper[4715]: I1204 14:07:38.758196 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:07:38 crc kubenswrapper[4715]: I1204 14:07:38.759208 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:08:08 crc kubenswrapper[4715]: I1204 14:08:08.758960 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:08:08 crc kubenswrapper[4715]: I1204 14:08:08.759615 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:08:08 crc kubenswrapper[4715]: I1204 14:08:08.759669 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:08:08 crc kubenswrapper[4715]: I1204 14:08:08.760391 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a6252ca8ea39b113b0e870cff0c9d7d47a3fe7fcf225413ba4908b3d3146389"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:08:08 crc kubenswrapper[4715]: I1204 14:08:08.760466 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://6a6252ca8ea39b113b0e870cff0c9d7d47a3fe7fcf225413ba4908b3d3146389" gracePeriod=600 Dec 04 14:08:09 crc kubenswrapper[4715]: I1204 14:08:09.679986 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="6a6252ca8ea39b113b0e870cff0c9d7d47a3fe7fcf225413ba4908b3d3146389" exitCode=0 Dec 04 14:08:09 crc kubenswrapper[4715]: I1204 14:08:09.680094 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"6a6252ca8ea39b113b0e870cff0c9d7d47a3fe7fcf225413ba4908b3d3146389"} Dec 04 14:08:09 crc kubenswrapper[4715]: I1204 14:08:09.680956 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"e37f663cd84616cad5630c875a686e056c8f65437b236f02818772405f217b98"} Dec 04 14:08:09 crc kubenswrapper[4715]: I1204 14:08:09.681021 4715 scope.go:117] "RemoveContainer" containerID="a679b434089a9d4f4ce6522120aa8528b8358549b266643a6b526d3fdfd66568" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.252560 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4tnrx"] Dec 04 14:10:11 crc kubenswrapper[4715]: E1204 14:10:11.253396 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678577cf-42b5-4f4a-84d2-9be8e1d7b753" containerName="registry" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.253414 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="678577cf-42b5-4f4a-84d2-9be8e1d7b753" containerName="registry" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.253730 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="678577cf-42b5-4f4a-84d2-9be8e1d7b753" containerName="registry" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.254209 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.256192 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.256289 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.256761 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-zx77c" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.259923 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-8nphx"] Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.260790 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-8nphx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.262933 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-b5p9q" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.272112 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4tnrx"] Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.286779 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qbkqd"] Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.288192 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.291650 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mqw9k" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.300190 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-8nphx"] Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.310802 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qbkqd"] Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.346783 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq9bd\" (UniqueName: \"kubernetes.io/projected/f4374824-e956-48c8-a4e5-52c55c7486e6-kube-api-access-bq9bd\") pod \"cert-manager-cainjector-7f985d654d-4tnrx\" (UID: \"f4374824-e956-48c8-a4e5-52c55c7486e6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.346893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr7ss\" (UniqueName: \"kubernetes.io/projected/70e6d48d-4b22-42ea-9016-a918b6889d78-kube-api-access-jr7ss\") pod \"cert-manager-5b446d88c5-8nphx\" (UID: \"70e6d48d-4b22-42ea-9016-a918b6889d78\") " pod="cert-manager/cert-manager-5b446d88c5-8nphx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.346930 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbsw\" (UniqueName: \"kubernetes.io/projected/288f92f5-abcf-40e4-937a-93f6f5ba002b-kube-api-access-blbsw\") pod \"cert-manager-webhook-5655c58dd6-qbkqd\" (UID: \"288f92f5-abcf-40e4-937a-93f6f5ba002b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.447287 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr7ss\" (UniqueName: \"kubernetes.io/projected/70e6d48d-4b22-42ea-9016-a918b6889d78-kube-api-access-jr7ss\") pod \"cert-manager-5b446d88c5-8nphx\" (UID: \"70e6d48d-4b22-42ea-9016-a918b6889d78\") " pod="cert-manager/cert-manager-5b446d88c5-8nphx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.447347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbsw\" (UniqueName: \"kubernetes.io/projected/288f92f5-abcf-40e4-937a-93f6f5ba002b-kube-api-access-blbsw\") pod \"cert-manager-webhook-5655c58dd6-qbkqd\" (UID: \"288f92f5-abcf-40e4-937a-93f6f5ba002b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.447372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq9bd\" (UniqueName: \"kubernetes.io/projected/f4374824-e956-48c8-a4e5-52c55c7486e6-kube-api-access-bq9bd\") pod \"cert-manager-cainjector-7f985d654d-4tnrx\" (UID: \"f4374824-e956-48c8-a4e5-52c55c7486e6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.464567 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr7ss\" (UniqueName: \"kubernetes.io/projected/70e6d48d-4b22-42ea-9016-a918b6889d78-kube-api-access-jr7ss\") pod \"cert-manager-5b446d88c5-8nphx\" (UID: \"70e6d48d-4b22-42ea-9016-a918b6889d78\") " pod="cert-manager/cert-manager-5b446d88c5-8nphx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.464786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbsw\" (UniqueName: \"kubernetes.io/projected/288f92f5-abcf-40e4-937a-93f6f5ba002b-kube-api-access-blbsw\") pod \"cert-manager-webhook-5655c58dd6-qbkqd\" (UID: \"288f92f5-abcf-40e4-937a-93f6f5ba002b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.466172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq9bd\" (UniqueName: \"kubernetes.io/projected/f4374824-e956-48c8-a4e5-52c55c7486e6-kube-api-access-bq9bd\") pod \"cert-manager-cainjector-7f985d654d-4tnrx\" (UID: \"f4374824-e956-48c8-a4e5-52c55c7486e6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.571866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.597882 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-8nphx" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.607439 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.875565 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qbkqd"] Dec 04 14:10:11 crc kubenswrapper[4715]: I1204 14:10:11.883112 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:10:12 crc kubenswrapper[4715]: I1204 14:10:12.011939 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4tnrx"] Dec 04 14:10:12 crc kubenswrapper[4715]: W1204 14:10:12.018548 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70e6d48d_4b22_42ea_9016_a918b6889d78.slice/crio-8692a30b09af97ecdb084c9b0abf4acf054cafa2399ce8e40733da6f20ea0b09 WatchSource:0}: Error finding container 8692a30b09af97ecdb084c9b0abf4acf054cafa2399ce8e40733da6f20ea0b09: Status 404 returned error can't find the container with id 8692a30b09af97ecdb084c9b0abf4acf054cafa2399ce8e40733da6f20ea0b09 Dec 04 14:10:12 crc kubenswrapper[4715]: I1204 14:10:12.018745 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-8nphx"] Dec 04 14:10:12 crc kubenswrapper[4715]: I1204 14:10:12.326320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" event={"ID":"f4374824-e956-48c8-a4e5-52c55c7486e6","Type":"ContainerStarted","Data":"05667aa7b1e264269cf52397c9f271db6bdb368b5e240b0b6937952093bd9540"} Dec 04 14:10:12 crc kubenswrapper[4715]: I1204 14:10:12.327909 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" event={"ID":"288f92f5-abcf-40e4-937a-93f6f5ba002b","Type":"ContainerStarted","Data":"2b8e158204da044e964ec8843339cb54e4dc704b419e640800c05d0acc58e87d"} Dec 04 14:10:12 crc kubenswrapper[4715]: I1204 14:10:12.328753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-8nphx" event={"ID":"70e6d48d-4b22-42ea-9016-a918b6889d78","Type":"ContainerStarted","Data":"8692a30b09af97ecdb084c9b0abf4acf054cafa2399ce8e40733da6f20ea0b09"} Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.359085 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" event={"ID":"f4374824-e956-48c8-a4e5-52c55c7486e6","Type":"ContainerStarted","Data":"37b6d595953697c3d52dd541f9a9cc579f1b52cb1c25f0de23ecc08e4b511098"} Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.364771 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" event={"ID":"288f92f5-abcf-40e4-937a-93f6f5ba002b","Type":"ContainerStarted","Data":"42c0fdbd5212eb1968ecf1fc3e8d8c9d58ec6d6c89ab801508427e1487e81b2f"} Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.365012 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.366574 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-8nphx" event={"ID":"70e6d48d-4b22-42ea-9016-a918b6889d78","Type":"ContainerStarted","Data":"5af3ea3c064f53db29da673163e4d5cee528de798ff49d2eb263f45ff792f022"} Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.401903 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-4tnrx" podStartSLOduration=1.787172749 podStartE2EDuration="5.401880774s" podCreationTimestamp="2025-12-04 14:10:11 +0000 UTC" firstStartedPulling="2025-12-04 14:10:12.019735006 +0000 UTC m=+769.088453221" lastFinishedPulling="2025-12-04 14:10:15.634443031 +0000 UTC m=+772.703161246" observedRunningTime="2025-12-04 14:10:16.385617079 +0000 UTC m=+773.454335324" watchObservedRunningTime="2025-12-04 14:10:16.401880774 +0000 UTC m=+773.470598989" Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.401999 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" podStartSLOduration=1.6429114139999998 podStartE2EDuration="5.401996378s" podCreationTimestamp="2025-12-04 14:10:11 +0000 UTC" firstStartedPulling="2025-12-04 14:10:11.882762141 +0000 UTC m=+768.951480356" lastFinishedPulling="2025-12-04 14:10:15.641847105 +0000 UTC m=+772.710565320" observedRunningTime="2025-12-04 14:10:16.397104369 +0000 UTC m=+773.465822584" watchObservedRunningTime="2025-12-04 14:10:16.401996378 +0000 UTC m=+773.470714593" Dec 04 14:10:16 crc kubenswrapper[4715]: I1204 14:10:16.413228 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-8nphx" podStartSLOduration=1.796452132 podStartE2EDuration="5.413206121s" podCreationTimestamp="2025-12-04 14:10:11 +0000 UTC" firstStartedPulling="2025-12-04 14:10:12.0199136 +0000 UTC m=+769.088631815" lastFinishedPulling="2025-12-04 14:10:15.636667589 +0000 UTC m=+772.705385804" observedRunningTime="2025-12-04 14:10:16.412230725 +0000 UTC m=+773.480948950" watchObservedRunningTime="2025-12-04 14:10:16.413206121 +0000 UTC m=+773.481924336" Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.609620 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbkqd" Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770293 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nw55d"] Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770712 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-controller" containerID="cri-o://40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770775 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="northd" containerID="cri-o://2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770857 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-acl-logging" containerID="cri-o://d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770816 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="sbdb" containerID="cri-o://6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770802 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="nbdb" containerID="cri-o://9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770837 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.770858 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-node" containerID="cri-o://9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" gracePeriod=30 Dec 04 14:10:21 crc kubenswrapper[4715]: I1204 14:10:21.840624 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" containerID="cri-o://301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" gracePeriod=30 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.122224 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/4.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.124351 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovn-acl-logging/0.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.124939 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovn-controller/0.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.125308 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.175919 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f49wk"] Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176168 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176185 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176199 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-node" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176206 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-node" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176216 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176225 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176239 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176246 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176257 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="northd" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176265 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="northd" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176275 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176282 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176291 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="nbdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176298 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="nbdb" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176308 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="sbdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176316 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="sbdb" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176328 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-acl-logging" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176335 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-acl-logging" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176348 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kubecfg-setup" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176355 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kubecfg-setup" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176368 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176376 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176388 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176395 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176506 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176520 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176528 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-acl-logging" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176538 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-node" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176547 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="northd" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176554 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176562 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176574 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="sbdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176585 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovn-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176595 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176603 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176613 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="nbdb" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176734 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176745 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.176759 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176766 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.176912 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerName="ovnkube-controller" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.178644 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182659 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-systemd-units\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182706 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182727 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovn-node-metrics-cert\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182770 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-var-lib-openvswitch\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182788 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-env-overrides\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-etc-openvswitch\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182852 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-node-log\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182844 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182888 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.182918 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183186 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-node-log" (OuterVolumeSpecName: "node-log") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183598 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183726 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-ovn\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183759 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183783 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-systemd\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183813 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-script-lib\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183829 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-log-socket\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183857 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xxk7\" (UniqueName: \"kubernetes.io/projected/727ec519-5a38-4de0-bc0f-69993f95c5ce-kube-api-access-9xxk7\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183878 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-config\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183885 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-log-socket" (OuterVolumeSpecName: "log-socket") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183903 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-bin\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183920 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-slash\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183938 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-kubelet\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183962 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-netd\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183988 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-openvswitch\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.183991 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-slash" (OuterVolumeSpecName: "host-slash") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184013 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-netns\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184025 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184054 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-ovn-kubernetes\") pod \"727ec519-5a38-4de0-bc0f-69993f95c5ce\" (UID: \"727ec519-5a38-4de0-bc0f-69993f95c5ce\") " Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184071 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184094 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184118 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184140 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184222 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184234 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovnkube-script-lib\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184250 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184275 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-log-socket\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184336 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-kubelet\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184414 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-run-ovn-kubernetes\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184445 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-node-log\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184479 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-ovn\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-cni-bin\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184545 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-systemd-units\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184595 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovn-node-metrics-cert\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184620 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-env-overrides\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184616 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184691 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wd5t\" (UniqueName: \"kubernetes.io/projected/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-kube-api-access-4wd5t\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184718 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-slash\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184802 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-run-netns\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184869 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-var-lib-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184901 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovnkube-config\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184947 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-cni-netd\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.184976 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-systemd\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185002 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-etc-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185100 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185116 4715 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185130 4715 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185143 4715 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185155 4715 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185165 4715 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185178 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185220 4715 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185232 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185243 4715 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185282 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185318 4715 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185332 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185343 4715 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185353 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185387 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.185399 4715 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.190450 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727ec519-5a38-4de0-bc0f-69993f95c5ce-kube-api-access-9xxk7" (OuterVolumeSpecName: "kube-api-access-9xxk7") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "kube-api-access-9xxk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.191272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.200259 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "727ec519-5a38-4de0-bc0f-69993f95c5ce" (UID: "727ec519-5a38-4de0-bc0f-69993f95c5ce"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286370 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovnkube-config\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286417 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-cni-netd\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-systemd\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286457 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-etc-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286488 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovnkube-script-lib\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286505 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-log-socket\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286521 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-kubelet\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286537 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-node-log\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-run-ovn-kubernetes\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286570 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-ovn\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-cni-bin\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286606 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-systemd-units\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovn-node-metrics-cert\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286670 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-env-overrides\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wd5t\" (UniqueName: \"kubernetes.io/projected/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-kube-api-access-4wd5t\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286704 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-slash\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-run-netns\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286818 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-var-lib-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-cni-netd\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286880 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-systemd\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-etc-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.286774 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-var-lib-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-systemd-units\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287186 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-cni-bin\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-run-netns\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287204 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-run-ovn-kubernetes\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287223 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-openvswitch\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287271 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-run-ovn\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287287 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-kubelet\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-node-log\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287309 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-host-slash\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287531 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-log-socket\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287622 4715 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/727ec519-5a38-4de0-bc0f-69993f95c5ce-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287631 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovnkube-script-lib\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287640 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xxk7\" (UniqueName: \"kubernetes.io/projected/727ec519-5a38-4de0-bc0f-69993f95c5ce-kube-api-access-9xxk7\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287660 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/727ec519-5a38-4de0-bc0f-69993f95c5ce-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287818 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovnkube-config\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.287867 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-env-overrides\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.291400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-ovn-node-metrics-cert\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.307016 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wd5t\" (UniqueName: \"kubernetes.io/projected/41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d-kube-api-access-4wd5t\") pod \"ovnkube-node-f49wk\" (UID: \"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d\") " pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.399244 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/3.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.399914 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/2.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.399965 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" containerID="d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1" exitCode=2 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.400051 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerDied","Data":"d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.400092 4715 scope.go:117] "RemoveContainer" containerID="25053101cdf55b00e9a430be0ba61e93577208d3729c5210863d55cd090e5ce6" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.400824 4715 scope.go:117] "RemoveContainer" containerID="d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.401144 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-multus pod=multus-8dlbk_openshift-multus(4ae4f7b7-e122-4bf8-b802-54e882e3d6cc)\"" pod="openshift-multus/multus-8dlbk" podUID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.404661 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovnkube-controller/4.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.406986 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovn-acl-logging/0.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407453 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nw55d_727ec519-5a38-4de0-bc0f-69993f95c5ce/ovn-controller/0.log" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407745 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" exitCode=0 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407766 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" exitCode=0 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407775 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" exitCode=0 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407769 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407814 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407831 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407782 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" exitCode=0 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407866 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" exitCode=0 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407879 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" exitCode=0 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407885 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407891 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" exitCode=143 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408085 4715 generic.go:334] "Generic (PLEG): container finished" podID="727ec519-5a38-4de0-bc0f-69993f95c5ce" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" exitCode=143 Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.407899 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408163 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408174 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408180 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408185 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408191 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408197 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408202 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408208 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408214 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408219 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408226 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408234 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408240 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408245 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408250 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408255 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408260 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408269 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408274 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408279 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408284 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408299 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408306 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408311 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408316 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408321 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408327 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408332 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408337 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408341 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408346 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408353 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nw55d" event={"ID":"727ec519-5a38-4de0-bc0f-69993f95c5ce","Type":"ContainerDied","Data":"20feddcaa9d707bb1925572f3b50d2844d39195328ac8c35cfc1808eecbddd4b"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408360 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408365 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408370 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408375 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408380 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408385 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408391 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408395 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408401 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.408406 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.427457 4715 scope.go:117] "RemoveContainer" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.443410 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nw55d"] Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.447895 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.448235 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nw55d"] Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.464590 4715 scope.go:117] "RemoveContainer" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.481349 4715 scope.go:117] "RemoveContainer" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.492267 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.496105 4715 scope.go:117] "RemoveContainer" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.510480 4715 scope.go:117] "RemoveContainer" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.521726 4715 scope.go:117] "RemoveContainer" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.537883 4715 scope.go:117] "RemoveContainer" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.549786 4715 scope.go:117] "RemoveContainer" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.575535 4715 scope.go:117] "RemoveContainer" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.607045 4715 scope.go:117] "RemoveContainer" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.608066 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": container with ID starting with 301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6 not found: ID does not exist" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.608110 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} err="failed to get container status \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": rpc error: code = NotFound desc = could not find container \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": container with ID starting with 301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.608141 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.608533 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": container with ID starting with ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a not found: ID does not exist" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.608562 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} err="failed to get container status \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": rpc error: code = NotFound desc = could not find container \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": container with ID starting with ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.608584 4715 scope.go:117] "RemoveContainer" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.608891 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": container with ID starting with 6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84 not found: ID does not exist" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.609029 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} err="failed to get container status \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": rpc error: code = NotFound desc = could not find container \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": container with ID starting with 6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.609122 4715 scope.go:117] "RemoveContainer" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.609487 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": container with ID starting with 9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725 not found: ID does not exist" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.609537 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} err="failed to get container status \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": rpc error: code = NotFound desc = could not find container \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": container with ID starting with 9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.609552 4715 scope.go:117] "RemoveContainer" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.609897 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": container with ID starting with 2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb not found: ID does not exist" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.609958 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} err="failed to get container status \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": rpc error: code = NotFound desc = could not find container \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": container with ID starting with 2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.609995 4715 scope.go:117] "RemoveContainer" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.610435 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": container with ID starting with 52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40 not found: ID does not exist" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.610500 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} err="failed to get container status \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": rpc error: code = NotFound desc = could not find container \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": container with ID starting with 52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.610516 4715 scope.go:117] "RemoveContainer" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.610955 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": container with ID starting with 9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794 not found: ID does not exist" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.610995 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} err="failed to get container status \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": rpc error: code = NotFound desc = could not find container \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": container with ID starting with 9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.611017 4715 scope.go:117] "RemoveContainer" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.611364 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": container with ID starting with d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20 not found: ID does not exist" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.611404 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} err="failed to get container status \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": rpc error: code = NotFound desc = could not find container \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": container with ID starting with d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.611456 4715 scope.go:117] "RemoveContainer" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.612343 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": container with ID starting with 40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca not found: ID does not exist" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.612380 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} err="failed to get container status \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": rpc error: code = NotFound desc = could not find container \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": container with ID starting with 40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.612402 4715 scope.go:117] "RemoveContainer" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" Dec 04 14:10:22 crc kubenswrapper[4715]: E1204 14:10:22.612841 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": container with ID starting with 1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c not found: ID does not exist" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.612881 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} err="failed to get container status \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": rpc error: code = NotFound desc = could not find container \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": container with ID starting with 1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.612908 4715 scope.go:117] "RemoveContainer" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.613338 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} err="failed to get container status \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": rpc error: code = NotFound desc = could not find container \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": container with ID starting with 301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.613368 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.613725 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} err="failed to get container status \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": rpc error: code = NotFound desc = could not find container \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": container with ID starting with ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.613761 4715 scope.go:117] "RemoveContainer" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.614101 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} err="failed to get container status \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": rpc error: code = NotFound desc = could not find container \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": container with ID starting with 6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.614192 4715 scope.go:117] "RemoveContainer" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.614562 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} err="failed to get container status \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": rpc error: code = NotFound desc = could not find container \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": container with ID starting with 9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.614594 4715 scope.go:117] "RemoveContainer" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.614945 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} err="failed to get container status \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": rpc error: code = NotFound desc = could not find container \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": container with ID starting with 2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.614976 4715 scope.go:117] "RemoveContainer" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.615278 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} err="failed to get container status \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": rpc error: code = NotFound desc = could not find container \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": container with ID starting with 52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.615297 4715 scope.go:117] "RemoveContainer" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.615590 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} err="failed to get container status \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": rpc error: code = NotFound desc = could not find container \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": container with ID starting with 9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.615607 4715 scope.go:117] "RemoveContainer" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.615810 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} err="failed to get container status \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": rpc error: code = NotFound desc = could not find container \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": container with ID starting with d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.615827 4715 scope.go:117] "RemoveContainer" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.616242 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} err="failed to get container status \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": rpc error: code = NotFound desc = could not find container \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": container with ID starting with 40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.616261 4715 scope.go:117] "RemoveContainer" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.616527 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} err="failed to get container status \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": rpc error: code = NotFound desc = could not find container \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": container with ID starting with 1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.616570 4715 scope.go:117] "RemoveContainer" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.617019 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} err="failed to get container status \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": rpc error: code = NotFound desc = could not find container \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": container with ID starting with 301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.617153 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.617591 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} err="failed to get container status \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": rpc error: code = NotFound desc = could not find container \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": container with ID starting with ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.617616 4715 scope.go:117] "RemoveContainer" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.618267 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} err="failed to get container status \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": rpc error: code = NotFound desc = could not find container \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": container with ID starting with 6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.618299 4715 scope.go:117] "RemoveContainer" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.618623 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} err="failed to get container status \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": rpc error: code = NotFound desc = could not find container \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": container with ID starting with 9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.618649 4715 scope.go:117] "RemoveContainer" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.618933 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} err="failed to get container status \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": rpc error: code = NotFound desc = could not find container \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": container with ID starting with 2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.618967 4715 scope.go:117] "RemoveContainer" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619150 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} err="failed to get container status \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": rpc error: code = NotFound desc = could not find container \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": container with ID starting with 52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619167 4715 scope.go:117] "RemoveContainer" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619295 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} err="failed to get container status \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": rpc error: code = NotFound desc = could not find container \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": container with ID starting with 9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619312 4715 scope.go:117] "RemoveContainer" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619457 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} err="failed to get container status \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": rpc error: code = NotFound desc = could not find container \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": container with ID starting with d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619480 4715 scope.go:117] "RemoveContainer" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619632 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} err="failed to get container status \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": rpc error: code = NotFound desc = could not find container \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": container with ID starting with 40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619652 4715 scope.go:117] "RemoveContainer" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619797 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} err="failed to get container status \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": rpc error: code = NotFound desc = could not find container \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": container with ID starting with 1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.619819 4715 scope.go:117] "RemoveContainer" containerID="301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.620083 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6"} err="failed to get container status \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": rpc error: code = NotFound desc = could not find container \"301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6\": container with ID starting with 301ed6665c7dc1a0f0c66aa3b15247fed2a674670f9147c6977b0ed8f574dcb6 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.620113 4715 scope.go:117] "RemoveContainer" containerID="ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.620733 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a"} err="failed to get container status \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": rpc error: code = NotFound desc = could not find container \"ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a\": container with ID starting with ac1518c65bf427a7cb9a4eca33b54be2c66ea9c593b729a983c909f169c3074a not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.620760 4715 scope.go:117] "RemoveContainer" containerID="6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.621179 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84"} err="failed to get container status \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": rpc error: code = NotFound desc = could not find container \"6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84\": container with ID starting with 6fde65a2b938b726ea944a6db38e266c5a48e774c3687d69459f24bd87ed8a84 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.621202 4715 scope.go:117] "RemoveContainer" containerID="9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.621539 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725"} err="failed to get container status \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": rpc error: code = NotFound desc = could not find container \"9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725\": container with ID starting with 9884c42ac726f0c5957e6916f2e95a427126b9dc45e7d5c412003573e1bd3725 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.621559 4715 scope.go:117] "RemoveContainer" containerID="2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.621912 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb"} err="failed to get container status \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": rpc error: code = NotFound desc = could not find container \"2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb\": container with ID starting with 2dd2b15b196797dfdddea1c3c7698fbeca591ccce9781d4271abe08877a87cdb not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.621933 4715 scope.go:117] "RemoveContainer" containerID="52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.622338 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40"} err="failed to get container status \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": rpc error: code = NotFound desc = could not find container \"52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40\": container with ID starting with 52d4c6b315d174464692d59cbde0bbc78673a15a6019fec6ed8568bfc0d68a40 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.622358 4715 scope.go:117] "RemoveContainer" containerID="9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.622606 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794"} err="failed to get container status \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": rpc error: code = NotFound desc = could not find container \"9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794\": container with ID starting with 9e7f7fd584df3dc996d6688ec8600ba7b35a922982e23112c2e42261a8f5b794 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.622627 4715 scope.go:117] "RemoveContainer" containerID="d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.622920 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20"} err="failed to get container status \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": rpc error: code = NotFound desc = could not find container \"d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20\": container with ID starting with d0117bd9dfcdbfdee82f9f9bb154140a31e72ac5fd3a1d52395c5acd35ae2e20 not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.622947 4715 scope.go:117] "RemoveContainer" containerID="40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.623315 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca"} err="failed to get container status \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": rpc error: code = NotFound desc = could not find container \"40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca\": container with ID starting with 40dd460f0b1d09c2fb84619ce3255e75a57feed368440d926fda2c6396a32eca not found: ID does not exist" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.623344 4715 scope.go:117] "RemoveContainer" containerID="1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c" Dec 04 14:10:22 crc kubenswrapper[4715]: I1204 14:10:22.623646 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c"} err="failed to get container status \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": rpc error: code = NotFound desc = could not find container \"1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c\": container with ID starting with 1ae701ee258e3c8b9e039716f0be82c78a65b05e8dd0b6107741f7f8963dbf6c not found: ID does not exist" Dec 04 14:10:23 crc kubenswrapper[4715]: I1204 14:10:23.188682 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="727ec519-5a38-4de0-bc0f-69993f95c5ce" path="/var/lib/kubelet/pods/727ec519-5a38-4de0-bc0f-69993f95c5ce/volumes" Dec 04 14:10:23 crc kubenswrapper[4715]: I1204 14:10:23.414210 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/3.log" Dec 04 14:10:23 crc kubenswrapper[4715]: I1204 14:10:23.415769 4715 generic.go:334] "Generic (PLEG): container finished" podID="41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d" containerID="22ce179d05bc2e2e250db1733a7bc6f73258ff0d6e62c60a6892d4aa287028f8" exitCode=0 Dec 04 14:10:23 crc kubenswrapper[4715]: I1204 14:10:23.415811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerDied","Data":"22ce179d05bc2e2e250db1733a7bc6f73258ff0d6e62c60a6892d4aa287028f8"} Dec 04 14:10:23 crc kubenswrapper[4715]: I1204 14:10:23.415836 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"504b4aedcaa20e84d30ee5e1766208efd4026260e91790c86b6b0f8f4e74e2d8"} Dec 04 14:10:24 crc kubenswrapper[4715]: I1204 14:10:24.433696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"e291627869ac9c43fdd06b3565ead262746110e60f8e9bc6d8f1490cc600652c"} Dec 04 14:10:24 crc kubenswrapper[4715]: I1204 14:10:24.434060 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"d1b8fcd8e42fc3cbf2d78869133b658a8bda25cfd38f558077a3a98f467618f7"} Dec 04 14:10:24 crc kubenswrapper[4715]: I1204 14:10:24.434077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"74fbf0c4d033730b552d477462a707b44541e2b12c471e564902cd8a2014f1a1"} Dec 04 14:10:24 crc kubenswrapper[4715]: I1204 14:10:24.434090 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"7b84df432bd1a10e1b84291867631078ae013ed929f2b6d243d505d90d125069"} Dec 04 14:10:24 crc kubenswrapper[4715]: I1204 14:10:24.434100 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"f2c6bfbd9caeb5832a13a45c1c39f0a9f43aab9c0316333f0629361ba1630eea"} Dec 04 14:10:24 crc kubenswrapper[4715]: I1204 14:10:24.434132 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"e1ccbee0d63316b7af6dade3125560e3bbf94156047ebebe3c55628cf403021a"} Dec 04 14:10:26 crc kubenswrapper[4715]: I1204 14:10:26.449024 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"1bfc67767f4eca8235fcac5fda8d815cb4c3112170e212e1ee09db9c63a54d4a"} Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.475753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" event={"ID":"41c1292c-ecfc-4ac6-bdcd-9b46d5dcc08d","Type":"ContainerStarted","Data":"189a5bec5479a5509ec1f56faa6fbb1412608257173db420803aaa44f366a65e"} Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.477726 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.478089 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.478275 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.515248 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.516985 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" podStartSLOduration=7.51696716 podStartE2EDuration="7.51696716s" podCreationTimestamp="2025-12-04 14:10:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:10:29.515578863 +0000 UTC m=+786.584297078" watchObservedRunningTime="2025-12-04 14:10:29.51696716 +0000 UTC m=+786.585685365" Dec 04 14:10:29 crc kubenswrapper[4715]: I1204 14:10:29.521573 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:10:35 crc kubenswrapper[4715]: I1204 14:10:35.180921 4715 scope.go:117] "RemoveContainer" containerID="d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1" Dec 04 14:10:35 crc kubenswrapper[4715]: E1204 14:10:35.181733 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-multus pod=multus-8dlbk_openshift-multus(4ae4f7b7-e122-4bf8-b802-54e882e3d6cc)\"" pod="openshift-multus/multus-8dlbk" podUID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" Dec 04 14:10:38 crc kubenswrapper[4715]: I1204 14:10:38.758637 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:10:38 crc kubenswrapper[4715]: I1204 14:10:38.758701 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:10:46 crc kubenswrapper[4715]: I1204 14:10:46.181171 4715 scope.go:117] "RemoveContainer" containerID="d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1" Dec 04 14:10:46 crc kubenswrapper[4715]: E1204 14:10:46.183735 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-multus pod=multus-8dlbk_openshift-multus(4ae4f7b7-e122-4bf8-b802-54e882e3d6cc)\"" pod="openshift-multus/multus-8dlbk" podUID="4ae4f7b7-e122-4bf8-b802-54e882e3d6cc" Dec 04 14:10:52 crc kubenswrapper[4715]: I1204 14:10:52.527570 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f49wk" Dec 04 14:11:00 crc kubenswrapper[4715]: I1204 14:11:00.180548 4715 scope.go:117] "RemoveContainer" containerID="d9d0f04376fc1c7da88347dbad5073b3ac56858885f0cfe560988dcf3f0c4fa1" Dec 04 14:11:00 crc kubenswrapper[4715]: I1204 14:11:00.664236 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8dlbk_4ae4f7b7-e122-4bf8-b802-54e882e3d6cc/kube-multus/3.log" Dec 04 14:11:00 crc kubenswrapper[4715]: I1204 14:11:00.664605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8dlbk" event={"ID":"4ae4f7b7-e122-4bf8-b802-54e882e3d6cc","Type":"ContainerStarted","Data":"09922c485d216ba59ae92d1385c42ff0c84254c41563aa3b4bef3a0aedab4b5b"} Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.128006 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj"] Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.129663 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.132987 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.137765 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj"] Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.327732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.327804 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqfx4\" (UniqueName: \"kubernetes.io/projected/b91a857d-7b05-4998-a029-e7576e7de6c9-kube-api-access-lqfx4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.327828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.428569 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqfx4\" (UniqueName: \"kubernetes.io/projected/b91a857d-7b05-4998-a029-e7576e7de6c9-kube-api-access-lqfx4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.428612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.428669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.429083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.429294 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.456317 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqfx4\" (UniqueName: \"kubernetes.io/projected/b91a857d-7b05-4998-a029-e7576e7de6c9-kube-api-access-lqfx4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: I1204 14:11:01.541595 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: E1204 14:11:01.730952 4715 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(8b22c2ebf092e949de88f3ea5cdc68aa2f98265c58d61dcae3536a00d5d2a05f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:11:01 crc kubenswrapper[4715]: E1204 14:11:01.731147 4715 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(8b22c2ebf092e949de88f3ea5cdc68aa2f98265c58d61dcae3536a00d5d2a05f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: E1204 14:11:01.731250 4715 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(8b22c2ebf092e949de88f3ea5cdc68aa2f98265c58d61dcae3536a00d5d2a05f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:01 crc kubenswrapper[4715]: E1204 14:11:01.731963 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace(b91a857d-7b05-4998-a029-e7576e7de6c9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace(b91a857d-7b05-4998-a029-e7576e7de6c9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(8b22c2ebf092e949de88f3ea5cdc68aa2f98265c58d61dcae3536a00d5d2a05f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" Dec 04 14:11:02 crc kubenswrapper[4715]: I1204 14:11:02.678293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:02 crc kubenswrapper[4715]: I1204 14:11:02.679426 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:02 crc kubenswrapper[4715]: E1204 14:11:02.704101 4715 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(f7d8f5e152624b8bc4652e4d9c462fb266503d827640a4634c5fe8858c87a471): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:11:02 crc kubenswrapper[4715]: E1204 14:11:02.704166 4715 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(f7d8f5e152624b8bc4652e4d9c462fb266503d827640a4634c5fe8858c87a471): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:02 crc kubenswrapper[4715]: E1204 14:11:02.704197 4715 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(f7d8f5e152624b8bc4652e4d9c462fb266503d827640a4634c5fe8858c87a471): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:02 crc kubenswrapper[4715]: E1204 14:11:02.704263 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace(b91a857d-7b05-4998-a029-e7576e7de6c9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace(b91a857d-7b05-4998-a029-e7576e7de6c9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_openshift-marketplace_b91a857d-7b05-4998-a029-e7576e7de6c9_0(f7d8f5e152624b8bc4652e4d9c462fb266503d827640a4634c5fe8858c87a471): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" Dec 04 14:11:08 crc kubenswrapper[4715]: I1204 14:11:08.759014 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:11:08 crc kubenswrapper[4715]: I1204 14:11:08.759750 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:11:13 crc kubenswrapper[4715]: I1204 14:11:13.516539 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 14:11:15 crc kubenswrapper[4715]: I1204 14:11:15.180434 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:15 crc kubenswrapper[4715]: I1204 14:11:15.181270 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:15 crc kubenswrapper[4715]: I1204 14:11:15.350227 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj"] Dec 04 14:11:15 crc kubenswrapper[4715]: W1204 14:11:15.359053 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb91a857d_7b05_4998_a029_e7576e7de6c9.slice/crio-ccb20d09b1b143a9e710ca2c9d13e9ac60e64ba87f01ccf846194d4658866169 WatchSource:0}: Error finding container ccb20d09b1b143a9e710ca2c9d13e9ac60e64ba87f01ccf846194d4658866169: Status 404 returned error can't find the container with id ccb20d09b1b143a9e710ca2c9d13e9ac60e64ba87f01ccf846194d4658866169 Dec 04 14:11:15 crc kubenswrapper[4715]: I1204 14:11:15.752629 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" event={"ID":"b91a857d-7b05-4998-a029-e7576e7de6c9","Type":"ContainerStarted","Data":"abd8e512fa9e9e5702d6975667cfa7d657e54946045b9818e2a40f863653fa57"} Dec 04 14:11:15 crc kubenswrapper[4715]: I1204 14:11:15.753365 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" event={"ID":"b91a857d-7b05-4998-a029-e7576e7de6c9","Type":"ContainerStarted","Data":"ccb20d09b1b143a9e710ca2c9d13e9ac60e64ba87f01ccf846194d4658866169"} Dec 04 14:11:16 crc kubenswrapper[4715]: I1204 14:11:16.759215 4715 generic.go:334] "Generic (PLEG): container finished" podID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerID="abd8e512fa9e9e5702d6975667cfa7d657e54946045b9818e2a40f863653fa57" exitCode=0 Dec 04 14:11:16 crc kubenswrapper[4715]: I1204 14:11:16.759339 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" event={"ID":"b91a857d-7b05-4998-a029-e7576e7de6c9","Type":"ContainerDied","Data":"abd8e512fa9e9e5702d6975667cfa7d657e54946045b9818e2a40f863653fa57"} Dec 04 14:11:18 crc kubenswrapper[4715]: I1204 14:11:18.773871 4715 generic.go:334] "Generic (PLEG): container finished" podID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerID="de7db18332dbaf6b75f7c857df25d72ad50fcfa858f73f1e834dacd923354bb1" exitCode=0 Dec 04 14:11:18 crc kubenswrapper[4715]: I1204 14:11:18.773957 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" event={"ID":"b91a857d-7b05-4998-a029-e7576e7de6c9","Type":"ContainerDied","Data":"de7db18332dbaf6b75f7c857df25d72ad50fcfa858f73f1e834dacd923354bb1"} Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.083617 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xl7d9"] Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.084764 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.096390 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xl7d9"] Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.152186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9v75\" (UniqueName: \"kubernetes.io/projected/d14da151-2200-4d47-8f10-a97ee710b431-kube-api-access-x9v75\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.152227 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-utilities\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.152285 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-catalog-content\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.253915 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9v75\" (UniqueName: \"kubernetes.io/projected/d14da151-2200-4d47-8f10-a97ee710b431-kube-api-access-x9v75\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.254011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-utilities\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.254136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-catalog-content\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.254690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-utilities\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.254901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-catalog-content\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.287535 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9v75\" (UniqueName: \"kubernetes.io/projected/d14da151-2200-4d47-8f10-a97ee710b431-kube-api-access-x9v75\") pod \"redhat-operators-xl7d9\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.449251 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.751266 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xl7d9"] Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.783566 4715 generic.go:334] "Generic (PLEG): container finished" podID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerID="74c299c843ec0b2e255693cdb9a5ad5aa87b6b750c5d24db98b37164e11ea1eb" exitCode=0 Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.783633 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" event={"ID":"b91a857d-7b05-4998-a029-e7576e7de6c9","Type":"ContainerDied","Data":"74c299c843ec0b2e255693cdb9a5ad5aa87b6b750c5d24db98b37164e11ea1eb"} Dec 04 14:11:19 crc kubenswrapper[4715]: I1204 14:11:19.785909 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerStarted","Data":"08129bda150e7e8d793e012381187fffad672a0dac61f07198e3f65e7e02f44c"} Dec 04 14:11:20 crc kubenswrapper[4715]: I1204 14:11:20.792095 4715 generic.go:334] "Generic (PLEG): container finished" podID="d14da151-2200-4d47-8f10-a97ee710b431" containerID="1ead88dc348e073d3846f0c2da2656bc6d16043b5c8db69b1748e035c1f10e6b" exitCode=0 Dec 04 14:11:20 crc kubenswrapper[4715]: I1204 14:11:20.792183 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerDied","Data":"1ead88dc348e073d3846f0c2da2656bc6d16043b5c8db69b1748e035c1f10e6b"} Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.049756 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.199408 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-bundle\") pod \"b91a857d-7b05-4998-a029-e7576e7de6c9\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.199585 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-util\") pod \"b91a857d-7b05-4998-a029-e7576e7de6c9\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.199802 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqfx4\" (UniqueName: \"kubernetes.io/projected/b91a857d-7b05-4998-a029-e7576e7de6c9-kube-api-access-lqfx4\") pod \"b91a857d-7b05-4998-a029-e7576e7de6c9\" (UID: \"b91a857d-7b05-4998-a029-e7576e7de6c9\") " Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.199977 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-bundle" (OuterVolumeSpecName: "bundle") pod "b91a857d-7b05-4998-a029-e7576e7de6c9" (UID: "b91a857d-7b05-4998-a029-e7576e7de6c9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.200132 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.206215 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b91a857d-7b05-4998-a029-e7576e7de6c9-kube-api-access-lqfx4" (OuterVolumeSpecName: "kube-api-access-lqfx4") pod "b91a857d-7b05-4998-a029-e7576e7de6c9" (UID: "b91a857d-7b05-4998-a029-e7576e7de6c9"). InnerVolumeSpecName "kube-api-access-lqfx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.212374 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-util" (OuterVolumeSpecName: "util") pod "b91a857d-7b05-4998-a029-e7576e7de6c9" (UID: "b91a857d-7b05-4998-a029-e7576e7de6c9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.301390 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b91a857d-7b05-4998-a029-e7576e7de6c9-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.301436 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqfx4\" (UniqueName: \"kubernetes.io/projected/b91a857d-7b05-4998-a029-e7576e7de6c9-kube-api-access-lqfx4\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.803912 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" event={"ID":"b91a857d-7b05-4998-a029-e7576e7de6c9","Type":"ContainerDied","Data":"ccb20d09b1b143a9e710ca2c9d13e9ac60e64ba87f01ccf846194d4658866169"} Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.804933 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccb20d09b1b143a9e710ca2c9d13e9ac60e64ba87f01ccf846194d4658866169" Dec 04 14:11:21 crc kubenswrapper[4715]: I1204 14:11:21.804906 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj" Dec 04 14:11:22 crc kubenswrapper[4715]: I1204 14:11:22.809479 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerStarted","Data":"d03fe0365bb71fcfce052fd629b7334b524728383be6442ab96eea637d9bc495"} Dec 04 14:11:26 crc kubenswrapper[4715]: I1204 14:11:26.835262 4715 generic.go:334] "Generic (PLEG): container finished" podID="d14da151-2200-4d47-8f10-a97ee710b431" containerID="d03fe0365bb71fcfce052fd629b7334b524728383be6442ab96eea637d9bc495" exitCode=0 Dec 04 14:11:26 crc kubenswrapper[4715]: I1204 14:11:26.835350 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerDied","Data":"d03fe0365bb71fcfce052fd629b7334b524728383be6442ab96eea637d9bc495"} Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.801320 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh"] Dec 04 14:11:27 crc kubenswrapper[4715]: E1204 14:11:27.801579 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="extract" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.801592 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="extract" Dec 04 14:11:27 crc kubenswrapper[4715]: E1204 14:11:27.801611 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="pull" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.801617 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="pull" Dec 04 14:11:27 crc kubenswrapper[4715]: E1204 14:11:27.801627 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="util" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.801634 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="util" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.801728 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b91a857d-7b05-4998-a029-e7576e7de6c9" containerName="extract" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.802196 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.804229 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.804238 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.804559 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-987vx" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.805023 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh"] Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.897683 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxqfp\" (UniqueName: \"kubernetes.io/projected/48b345c0-a5f8-4ff5-b696-e147e5ec9753-kube-api-access-qxqfp\") pod \"nmstate-operator-5b5b58f5c8-8k4vh\" (UID: \"48b345c0-a5f8-4ff5-b696-e147e5ec9753\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" Dec 04 14:11:27 crc kubenswrapper[4715]: I1204 14:11:27.998721 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxqfp\" (UniqueName: \"kubernetes.io/projected/48b345c0-a5f8-4ff5-b696-e147e5ec9753-kube-api-access-qxqfp\") pod \"nmstate-operator-5b5b58f5c8-8k4vh\" (UID: \"48b345c0-a5f8-4ff5-b696-e147e5ec9753\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" Dec 04 14:11:28 crc kubenswrapper[4715]: I1204 14:11:28.016382 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxqfp\" (UniqueName: \"kubernetes.io/projected/48b345c0-a5f8-4ff5-b696-e147e5ec9753-kube-api-access-qxqfp\") pod \"nmstate-operator-5b5b58f5c8-8k4vh\" (UID: \"48b345c0-a5f8-4ff5-b696-e147e5ec9753\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" Dec 04 14:11:28 crc kubenswrapper[4715]: I1204 14:11:28.120564 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" Dec 04 14:11:28 crc kubenswrapper[4715]: I1204 14:11:28.666821 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh"] Dec 04 14:11:28 crc kubenswrapper[4715]: I1204 14:11:28.864801 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerStarted","Data":"a4cda8a0d42f1d1c24797fceb0f808473b4b5392b40f4931eb3527d79bfa9728"} Dec 04 14:11:28 crc kubenswrapper[4715]: I1204 14:11:28.865815 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" event={"ID":"48b345c0-a5f8-4ff5-b696-e147e5ec9753","Type":"ContainerStarted","Data":"915cc2ed5c456168b62cb2acc2e53c33959508ef0e239702946aef4cbca451db"} Dec 04 14:11:28 crc kubenswrapper[4715]: I1204 14:11:28.881966 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xl7d9" podStartSLOduration=2.2363428499999998 podStartE2EDuration="9.88194939s" podCreationTimestamp="2025-12-04 14:11:19 +0000 UTC" firstStartedPulling="2025-12-04 14:11:20.79344001 +0000 UTC m=+837.862158215" lastFinishedPulling="2025-12-04 14:11:28.43904654 +0000 UTC m=+845.507764755" observedRunningTime="2025-12-04 14:11:28.881115528 +0000 UTC m=+845.949833753" watchObservedRunningTime="2025-12-04 14:11:28.88194939 +0000 UTC m=+845.950667605" Dec 04 14:11:29 crc kubenswrapper[4715]: I1204 14:11:29.450211 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:29 crc kubenswrapper[4715]: I1204 14:11:29.450520 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:30 crc kubenswrapper[4715]: I1204 14:11:30.514487 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xl7d9" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="registry-server" probeResult="failure" output=< Dec 04 14:11:30 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 04 14:11:30 crc kubenswrapper[4715]: > Dec 04 14:11:32 crc kubenswrapper[4715]: I1204 14:11:32.885295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" event={"ID":"48b345c0-a5f8-4ff5-b696-e147e5ec9753","Type":"ContainerStarted","Data":"2dabccaadf192cbd0a4355b5985126bb90b9956d0d41b3fc22277601c9043002"} Dec 04 14:11:32 crc kubenswrapper[4715]: I1204 14:11:32.899412 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8k4vh" podStartSLOduration=2.781961564 podStartE2EDuration="5.89939515s" podCreationTimestamp="2025-12-04 14:11:27 +0000 UTC" firstStartedPulling="2025-12-04 14:11:28.678970408 +0000 UTC m=+845.747688633" lastFinishedPulling="2025-12-04 14:11:31.796404004 +0000 UTC m=+848.865122219" observedRunningTime="2025-12-04 14:11:32.8968029 +0000 UTC m=+849.965521135" watchObservedRunningTime="2025-12-04 14:11:32.89939515 +0000 UTC m=+849.968113365" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.802094 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.803143 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.804993 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-w4vhq" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.816239 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.820412 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.821388 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.823102 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.844599 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-lml75"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.845498 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.891018 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.945637 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.946320 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.951355 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.951435 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-j54bk" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.951471 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.954741 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q"] Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-nmstate-lock\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-ovs-socket\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwvjm\" (UniqueName: \"kubernetes.io/projected/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-kube-api-access-gwvjm\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955687 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znw6s\" (UniqueName: \"kubernetes.io/projected/4165e1e3-7227-4666-a737-14083e64195d-kube-api-access-znw6s\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955761 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-dbus-socket\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:33 crc kubenswrapper[4715]: I1204 14:11:33.955831 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzr9g\" (UniqueName: \"kubernetes.io/projected/c4ab73a9-a74b-43dc-bbe5-ad71120feab2-kube-api-access-lzr9g\") pod \"nmstate-metrics-7f946cbc9-nc82h\" (UID: \"c4ab73a9-a74b-43dc-bbe5-ad71120feab2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znw6s\" (UniqueName: \"kubernetes.io/projected/4165e1e3-7227-4666-a737-14083e64195d-kube-api-access-znw6s\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057245 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057268 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-dbus-socket\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkk7q\" (UniqueName: \"kubernetes.io/projected/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-kube-api-access-kkk7q\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzr9g\" (UniqueName: \"kubernetes.io/projected/c4ab73a9-a74b-43dc-bbe5-ad71120feab2-kube-api-access-lzr9g\") pod \"nmstate-metrics-7f946cbc9-nc82h\" (UID: \"c4ab73a9-a74b-43dc-bbe5-ad71120feab2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057354 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-nmstate-lock\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057394 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-ovs-socket\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057435 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057457 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwvjm\" (UniqueName: \"kubernetes.io/projected/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-kube-api-access-gwvjm\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.057480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:34 crc kubenswrapper[4715]: E1204 14:11:34.057630 4715 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 04 14:11:34 crc kubenswrapper[4715]: E1204 14:11:34.057688 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-tls-key-pair podName:746b0b9b-20ec-42d8-8d9a-4aa11c18a58e nodeName:}" failed. No retries permitted until 2025-12-04 14:11:34.55766757 +0000 UTC m=+851.626385785 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-dbfzg" (UID: "746b0b9b-20ec-42d8-8d9a-4aa11c18a58e") : secret "openshift-nmstate-webhook" not found Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.058342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-dbus-socket\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.058550 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-nmstate-lock\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.058607 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4165e1e3-7227-4666-a737-14083e64195d-ovs-socket\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.086282 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzr9g\" (UniqueName: \"kubernetes.io/projected/c4ab73a9-a74b-43dc-bbe5-ad71120feab2-kube-api-access-lzr9g\") pod \"nmstate-metrics-7f946cbc9-nc82h\" (UID: \"c4ab73a9-a74b-43dc-bbe5-ad71120feab2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.087077 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwvjm\" (UniqueName: \"kubernetes.io/projected/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-kube-api-access-gwvjm\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.090587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znw6s\" (UniqueName: \"kubernetes.io/projected/4165e1e3-7227-4666-a737-14083e64195d-kube-api-access-znw6s\") pod \"nmstate-handler-lml75\" (UID: \"4165e1e3-7227-4666-a737-14083e64195d\") " pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.120817 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.158867 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.159575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.159621 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkk7q\" (UniqueName: \"kubernetes.io/projected/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-kube-api-access-kkk7q\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.161466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: E1204 14:11:34.161581 4715 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 04 14:11:34 crc kubenswrapper[4715]: E1204 14:11:34.161647 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-plugin-serving-cert podName:29cbcfcf-ff1d-4605-ac6e-d7c609de5911 nodeName:}" failed. No retries permitted until 2025-12-04 14:11:34.661629067 +0000 UTC m=+851.730347282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-pg74q" (UID: "29cbcfcf-ff1d-4605-ac6e-d7c609de5911") : secret "plugin-serving-cert" not found Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.178366 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.179341 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkk7q\" (UniqueName: \"kubernetes.io/projected/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-kube-api-access-kkk7q\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.391204 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7c589d8dc4-nhrrm"] Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.391959 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.448167 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c589d8dc4-nhrrm"] Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466616 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/763a22ef-b057-4159-bf3a-220825dd5d70-console-serving-cert\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466724 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-service-ca\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466756 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-console-config\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466805 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-oauth-serving-cert\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466842 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/763a22ef-b057-4159-bf3a-220825dd5d70-console-oauth-config\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466889 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-trusted-ca-bundle\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.466918 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kscd5\" (UniqueName: \"kubernetes.io/projected/763a22ef-b057-4159-bf3a-220825dd5d70-kube-api-access-kscd5\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.499694 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h"] Dec 04 14:11:34 crc kubenswrapper[4715]: W1204 14:11:34.508526 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4ab73a9_a74b_43dc_bbe5_ad71120feab2.slice/crio-e34ec670fde73ac26468ea5d4f128c11f6724456fac2529c1a7852bf0c3d0d81 WatchSource:0}: Error finding container e34ec670fde73ac26468ea5d4f128c11f6724456fac2529c1a7852bf0c3d0d81: Status 404 returned error can't find the container with id e34ec670fde73ac26468ea5d4f128c11f6724456fac2529c1a7852bf0c3d0d81 Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.568377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-trusted-ca-bundle\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.568441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kscd5\" (UniqueName: \"kubernetes.io/projected/763a22ef-b057-4159-bf3a-220825dd5d70-kube-api-access-kscd5\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.568475 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/763a22ef-b057-4159-bf3a-220825dd5d70-console-serving-cert\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.568539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.569440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-service-ca\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.569471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-console-config\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.569513 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-oauth-serving-cert\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.569549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/763a22ef-b057-4159-bf3a-220825dd5d70-console-oauth-config\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.570340 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-trusted-ca-bundle\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.570400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-console-config\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.570589 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-service-ca\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.570959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/763a22ef-b057-4159-bf3a-220825dd5d70-oauth-serving-cert\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.573574 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/763a22ef-b057-4159-bf3a-220825dd5d70-console-serving-cert\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.574055 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/763a22ef-b057-4159-bf3a-220825dd5d70-console-oauth-config\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.575461 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/746b0b9b-20ec-42d8-8d9a-4aa11c18a58e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dbfzg\" (UID: \"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.590415 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kscd5\" (UniqueName: \"kubernetes.io/projected/763a22ef-b057-4159-bf3a-220825dd5d70-kube-api-access-kscd5\") pod \"console-7c589d8dc4-nhrrm\" (UID: \"763a22ef-b057-4159-bf3a-220825dd5d70\") " pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.671010 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.676623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/29cbcfcf-ff1d-4605-ac6e-d7c609de5911-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-pg74q\" (UID: \"29cbcfcf-ff1d-4605-ac6e-d7c609de5911\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.740343 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.782938 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.862640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.899421 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lml75" event={"ID":"4165e1e3-7227-4666-a737-14083e64195d","Type":"ContainerStarted","Data":"e6ce4e4aa5d96cdf8d4ceef0094b3148af1e0f14da2139505fb7d509e2875580"} Dec 04 14:11:34 crc kubenswrapper[4715]: I1204 14:11:34.906102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" event={"ID":"c4ab73a9-a74b-43dc-bbe5-ad71120feab2","Type":"ContainerStarted","Data":"e34ec670fde73ac26468ea5d4f128c11f6724456fac2529c1a7852bf0c3d0d81"} Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.238099 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg"] Dec 04 14:11:35 crc kubenswrapper[4715]: W1204 14:11:35.268398 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod746b0b9b_20ec_42d8_8d9a_4aa11c18a58e.slice/crio-9b1dad99eb096c2d77b3b2d4a7e2fe82e6c0ffc6adc3347ee54e08d3eb1fe64e WatchSource:0}: Error finding container 9b1dad99eb096c2d77b3b2d4a7e2fe82e6c0ffc6adc3347ee54e08d3eb1fe64e: Status 404 returned error can't find the container with id 9b1dad99eb096c2d77b3b2d4a7e2fe82e6c0ffc6adc3347ee54e08d3eb1fe64e Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.529840 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c589d8dc4-nhrrm"] Dec 04 14:11:35 crc kubenswrapper[4715]: W1204 14:11:35.539388 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod763a22ef_b057_4159_bf3a_220825dd5d70.slice/crio-6b52cd05d628012d1bf15085354b492b75eb1810c5a457294f8128c95d9c6984 WatchSource:0}: Error finding container 6b52cd05d628012d1bf15085354b492b75eb1810c5a457294f8128c95d9c6984: Status 404 returned error can't find the container with id 6b52cd05d628012d1bf15085354b492b75eb1810c5a457294f8128c95d9c6984 Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.592556 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q"] Dec 04 14:11:35 crc kubenswrapper[4715]: W1204 14:11:35.596541 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29cbcfcf_ff1d_4605_ac6e_d7c609de5911.slice/crio-f01142f7aa35c070d5e95550a27be000a49223e9e042d0231caca6bbf7384ae9 WatchSource:0}: Error finding container f01142f7aa35c070d5e95550a27be000a49223e9e042d0231caca6bbf7384ae9: Status 404 returned error can't find the container with id f01142f7aa35c070d5e95550a27be000a49223e9e042d0231caca6bbf7384ae9 Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.923420 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c589d8dc4-nhrrm" event={"ID":"763a22ef-b057-4159-bf3a-220825dd5d70","Type":"ContainerStarted","Data":"0cd8149aba47dabcd4c52ace6f995b1aefdb7ba74c689503d46d3b8a46d5fa85"} Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.923754 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c589d8dc4-nhrrm" event={"ID":"763a22ef-b057-4159-bf3a-220825dd5d70","Type":"ContainerStarted","Data":"6b52cd05d628012d1bf15085354b492b75eb1810c5a457294f8128c95d9c6984"} Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.924732 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" event={"ID":"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e","Type":"ContainerStarted","Data":"9b1dad99eb096c2d77b3b2d4a7e2fe82e6c0ffc6adc3347ee54e08d3eb1fe64e"} Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.926174 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" event={"ID":"29cbcfcf-ff1d-4605-ac6e-d7c609de5911","Type":"ContainerStarted","Data":"f01142f7aa35c070d5e95550a27be000a49223e9e042d0231caca6bbf7384ae9"} Dec 04 14:11:35 crc kubenswrapper[4715]: I1204 14:11:35.942069 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7c589d8dc4-nhrrm" podStartSLOduration=1.942023282 podStartE2EDuration="1.942023282s" podCreationTimestamp="2025-12-04 14:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:35.938618692 +0000 UTC m=+853.007336917" watchObservedRunningTime="2025-12-04 14:11:35.942023282 +0000 UTC m=+853.010741497" Dec 04 14:11:38 crc kubenswrapper[4715]: I1204 14:11:38.758777 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:11:38 crc kubenswrapper[4715]: I1204 14:11:38.759281 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:11:38 crc kubenswrapper[4715]: I1204 14:11:38.759336 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:11:38 crc kubenswrapper[4715]: I1204 14:11:38.759939 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e37f663cd84616cad5630c875a686e056c8f65437b236f02818772405f217b98"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:11:38 crc kubenswrapper[4715]: I1204 14:11:38.760008 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://e37f663cd84616cad5630c875a686e056c8f65437b236f02818772405f217b98" gracePeriod=600 Dec 04 14:11:39 crc kubenswrapper[4715]: I1204 14:11:39.138049 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="e37f663cd84616cad5630c875a686e056c8f65437b236f02818772405f217b98" exitCode=0 Dec 04 14:11:39 crc kubenswrapper[4715]: I1204 14:11:39.138067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"e37f663cd84616cad5630c875a686e056c8f65437b236f02818772405f217b98"} Dec 04 14:11:39 crc kubenswrapper[4715]: I1204 14:11:39.138463 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"60dea1351c777fadc6f4bbc06b373dac4289c7f94189506529170a4446c48e46"} Dec 04 14:11:39 crc kubenswrapper[4715]: I1204 14:11:39.138492 4715 scope.go:117] "RemoveContainer" containerID="6a6252ca8ea39b113b0e870cff0c9d7d47a3fe7fcf225413ba4908b3d3146389" Dec 04 14:11:39 crc kubenswrapper[4715]: I1204 14:11:39.506149 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:39 crc kubenswrapper[4715]: I1204 14:11:39.562805 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.146538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lml75" event={"ID":"4165e1e3-7227-4666-a737-14083e64195d","Type":"ContainerStarted","Data":"d80ae2b3ab9aa45d035795e02c42ec1453dd7c813ccbc2d63b478bbea8269fce"} Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.146913 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.148515 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" event={"ID":"c4ab73a9-a74b-43dc-bbe5-ad71120feab2","Type":"ContainerStarted","Data":"b660e8b20b661e48fcf8142fe8fb8eec2552a2e4b5a0f1c6829950a14a6e2db2"} Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.153655 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" event={"ID":"746b0b9b-20ec-42d8-8d9a-4aa11c18a58e","Type":"ContainerStarted","Data":"d9e0b6329d666e7b14e66b4a5e6ed3f0ff73b4314f6191f0948ce44299c4e2ee"} Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.153699 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.167639 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-lml75" podStartSLOduration=2.296826573 podStartE2EDuration="7.167618087s" podCreationTimestamp="2025-12-04 14:11:33 +0000 UTC" firstStartedPulling="2025-12-04 14:11:34.199006175 +0000 UTC m=+851.267724390" lastFinishedPulling="2025-12-04 14:11:39.069797689 +0000 UTC m=+856.138515904" observedRunningTime="2025-12-04 14:11:40.165300865 +0000 UTC m=+857.234019080" watchObservedRunningTime="2025-12-04 14:11:40.167618087 +0000 UTC m=+857.236336312" Dec 04 14:11:40 crc kubenswrapper[4715]: I1204 14:11:40.196394 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" podStartSLOduration=3.39723867 podStartE2EDuration="7.196348914s" podCreationTimestamp="2025-12-04 14:11:33 +0000 UTC" firstStartedPulling="2025-12-04 14:11:35.272152034 +0000 UTC m=+852.340870249" lastFinishedPulling="2025-12-04 14:11:39.071262278 +0000 UTC m=+856.139980493" observedRunningTime="2025-12-04 14:11:40.182145165 +0000 UTC m=+857.250863380" watchObservedRunningTime="2025-12-04 14:11:40.196348914 +0000 UTC m=+857.265067129" Dec 04 14:11:41 crc kubenswrapper[4715]: I1204 14:11:41.162193 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" event={"ID":"29cbcfcf-ff1d-4605-ac6e-d7c609de5911","Type":"ContainerStarted","Data":"76d8ea6ecf8601da9e424239fd03399782a513e95906105023403cb5c37b03c1"} Dec 04 14:11:41 crc kubenswrapper[4715]: I1204 14:11:41.175844 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-pg74q" podStartSLOduration=3.629551653 podStartE2EDuration="8.17582542s" podCreationTimestamp="2025-12-04 14:11:33 +0000 UTC" firstStartedPulling="2025-12-04 14:11:35.599581408 +0000 UTC m=+852.668299623" lastFinishedPulling="2025-12-04 14:11:40.145855175 +0000 UTC m=+857.214573390" observedRunningTime="2025-12-04 14:11:41.17471074 +0000 UTC m=+858.243428955" watchObservedRunningTime="2025-12-04 14:11:41.17582542 +0000 UTC m=+858.244543645" Dec 04 14:11:41 crc kubenswrapper[4715]: I1204 14:11:41.903090 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xl7d9"] Dec 04 14:11:41 crc kubenswrapper[4715]: I1204 14:11:41.903704 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xl7d9" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="registry-server" containerID="cri-o://a4cda8a0d42f1d1c24797fceb0f808473b4b5392b40f4931eb3527d79bfa9728" gracePeriod=2 Dec 04 14:11:41 crc kubenswrapper[4715]: E1204 14:11:41.996760 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/systemd-tmpfiles-clean.service\": RecentStats: unable to find data in memory cache]" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.168868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" event={"ID":"c4ab73a9-a74b-43dc-bbe5-ad71120feab2","Type":"ContainerStarted","Data":"c08c69220570134704495cff323c8f0d293d74d614b81a4df79cf1e5fcc620d7"} Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.172132 4715 generic.go:334] "Generic (PLEG): container finished" podID="d14da151-2200-4d47-8f10-a97ee710b431" containerID="a4cda8a0d42f1d1c24797fceb0f808473b4b5392b40f4931eb3527d79bfa9728" exitCode=0 Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.172515 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerDied","Data":"a4cda8a0d42f1d1c24797fceb0f808473b4b5392b40f4931eb3527d79bfa9728"} Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.194229 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-nc82h" podStartSLOduration=2.008896494 podStartE2EDuration="9.194209736s" podCreationTimestamp="2025-12-04 14:11:33 +0000 UTC" firstStartedPulling="2025-12-04 14:11:34.512297702 +0000 UTC m=+851.581015917" lastFinishedPulling="2025-12-04 14:11:41.697610944 +0000 UTC m=+858.766329159" observedRunningTime="2025-12-04 14:11:42.1928712 +0000 UTC m=+859.261589415" watchObservedRunningTime="2025-12-04 14:11:42.194209736 +0000 UTC m=+859.262927951" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.347651 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.470551 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9v75\" (UniqueName: \"kubernetes.io/projected/d14da151-2200-4d47-8f10-a97ee710b431-kube-api-access-x9v75\") pod \"d14da151-2200-4d47-8f10-a97ee710b431\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.470722 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-utilities\") pod \"d14da151-2200-4d47-8f10-a97ee710b431\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.470757 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-catalog-content\") pod \"d14da151-2200-4d47-8f10-a97ee710b431\" (UID: \"d14da151-2200-4d47-8f10-a97ee710b431\") " Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.472115 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-utilities" (OuterVolumeSpecName: "utilities") pod "d14da151-2200-4d47-8f10-a97ee710b431" (UID: "d14da151-2200-4d47-8f10-a97ee710b431"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.475872 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d14da151-2200-4d47-8f10-a97ee710b431-kube-api-access-x9v75" (OuterVolumeSpecName: "kube-api-access-x9v75") pod "d14da151-2200-4d47-8f10-a97ee710b431" (UID: "d14da151-2200-4d47-8f10-a97ee710b431"). InnerVolumeSpecName "kube-api-access-x9v75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.572287 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.572343 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9v75\" (UniqueName: \"kubernetes.io/projected/d14da151-2200-4d47-8f10-a97ee710b431-kube-api-access-x9v75\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.574645 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d14da151-2200-4d47-8f10-a97ee710b431" (UID: "d14da151-2200-4d47-8f10-a97ee710b431"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:42 crc kubenswrapper[4715]: I1204 14:11:42.673712 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14da151-2200-4d47-8f10-a97ee710b431-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.182798 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xl7d9" Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.188267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xl7d9" event={"ID":"d14da151-2200-4d47-8f10-a97ee710b431","Type":"ContainerDied","Data":"08129bda150e7e8d793e012381187fffad672a0dac61f07198e3f65e7e02f44c"} Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.188312 4715 scope.go:117] "RemoveContainer" containerID="a4cda8a0d42f1d1c24797fceb0f808473b4b5392b40f4931eb3527d79bfa9728" Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.207429 4715 scope.go:117] "RemoveContainer" containerID="d03fe0365bb71fcfce052fd629b7334b524728383be6442ab96eea637d9bc495" Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.233315 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xl7d9"] Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.239429 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xl7d9"] Dec 04 14:11:43 crc kubenswrapper[4715]: I1204 14:11:43.244267 4715 scope.go:117] "RemoveContainer" containerID="1ead88dc348e073d3846f0c2da2656bc6d16043b5c8db69b1748e035c1f10e6b" Dec 04 14:11:44 crc kubenswrapper[4715]: I1204 14:11:44.208109 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-lml75" Dec 04 14:11:44 crc kubenswrapper[4715]: I1204 14:11:44.783945 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:44 crc kubenswrapper[4715]: I1204 14:11:44.784240 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:44 crc kubenswrapper[4715]: I1204 14:11:44.788617 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:45 crc kubenswrapper[4715]: I1204 14:11:45.188817 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d14da151-2200-4d47-8f10-a97ee710b431" path="/var/lib/kubelet/pods/d14da151-2200-4d47-8f10-a97ee710b431/volumes" Dec 04 14:11:45 crc kubenswrapper[4715]: I1204 14:11:45.199993 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7c589d8dc4-nhrrm" Dec 04 14:11:45 crc kubenswrapper[4715]: I1204 14:11:45.244464 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vvwbg"] Dec 04 14:11:54 crc kubenswrapper[4715]: I1204 14:11:54.746459 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dbfzg" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.446624 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s"] Dec 04 14:12:07 crc kubenswrapper[4715]: E1204 14:12:07.447268 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="registry-server" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.447281 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="registry-server" Dec 04 14:12:07 crc kubenswrapper[4715]: E1204 14:12:07.447296 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="extract-utilities" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.447302 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="extract-utilities" Dec 04 14:12:07 crc kubenswrapper[4715]: E1204 14:12:07.447311 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="extract-content" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.447318 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="extract-content" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.447418 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d14da151-2200-4d47-8f10-a97ee710b431" containerName="registry-server" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.448266 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.453304 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.474426 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s"] Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.638988 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.639813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88dxf\" (UniqueName: \"kubernetes.io/projected/e7c6ddac-1848-4449-8fbe-69d10990b25a-kube-api-access-88dxf\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.640008 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.741281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.741369 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.741401 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88dxf\" (UniqueName: \"kubernetes.io/projected/e7c6ddac-1848-4449-8fbe-69d10990b25a-kube-api-access-88dxf\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.741796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.741821 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.761255 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88dxf\" (UniqueName: \"kubernetes.io/projected/e7c6ddac-1848-4449-8fbe-69d10990b25a-kube-api-access-88dxf\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:07 crc kubenswrapper[4715]: I1204 14:12:07.778397 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:08 crc kubenswrapper[4715]: I1204 14:12:08.168428 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s"] Dec 04 14:12:08 crc kubenswrapper[4715]: I1204 14:12:08.357703 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" event={"ID":"e7c6ddac-1848-4449-8fbe-69d10990b25a","Type":"ContainerStarted","Data":"035a8a4fc708ca75486febc0144c614303df27b3598b5dfe1d76ba8fff71dbb1"} Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.283411 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vvwbg" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerName="console" containerID="cri-o://6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40" gracePeriod=15 Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.369104 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerID="a0bd63daf328d3b26c63af1d2a64448d5ad218414d93e49e592051b9a645987b" exitCode=0 Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.369167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" event={"ID":"e7c6ddac-1848-4449-8fbe-69d10990b25a","Type":"ContainerDied","Data":"a0bd63daf328d3b26c63af1d2a64448d5ad218414d93e49e592051b9a645987b"} Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.617076 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vvwbg_c3b08994-cb44-48a9-87d1-af33d2b511b6/console/0.log" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.617368 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.678344 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-serving-cert\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.678391 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-config\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.678434 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9h49\" (UniqueName: \"kubernetes.io/projected/c3b08994-cb44-48a9-87d1-af33d2b511b6-kube-api-access-n9h49\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.678453 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-oauth-config\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.678515 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-oauth-serving-cert\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679255 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-service-ca\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679356 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-trusted-ca-bundle\") pod \"c3b08994-cb44-48a9-87d1-af33d2b511b6\" (UID: \"c3b08994-cb44-48a9-87d1-af33d2b511b6\") " Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679383 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-service-ca" (OuterVolumeSpecName: "service-ca") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679666 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-config" (OuterVolumeSpecName: "console-config") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679821 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679841 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.679854 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.680003 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.685297 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.685595 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b08994-cb44-48a9-87d1-af33d2b511b6-kube-api-access-n9h49" (OuterVolumeSpecName: "kube-api-access-n9h49") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "kube-api-access-n9h49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.685651 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c3b08994-cb44-48a9-87d1-af33d2b511b6" (UID: "c3b08994-cb44-48a9-87d1-af33d2b511b6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.780603 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b08994-cb44-48a9-87d1-af33d2b511b6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.780648 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.780683 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9h49\" (UniqueName: \"kubernetes.io/projected/c3b08994-cb44-48a9-87d1-af33d2b511b6-kube-api-access-n9h49\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:10 crc kubenswrapper[4715]: I1204 14:12:10.780698 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3b08994-cb44-48a9-87d1-af33d2b511b6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.376320 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vvwbg_c3b08994-cb44-48a9-87d1-af33d2b511b6/console/0.log" Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.376383 4715 generic.go:334] "Generic (PLEG): container finished" podID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerID="6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40" exitCode=2 Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.376411 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvwbg" event={"ID":"c3b08994-cb44-48a9-87d1-af33d2b511b6","Type":"ContainerDied","Data":"6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40"} Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.376459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvwbg" event={"ID":"c3b08994-cb44-48a9-87d1-af33d2b511b6","Type":"ContainerDied","Data":"77d91947076b85cffb7ba970a4c5b6dc2156f3fe9f6f01e062f82681eebde3b5"} Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.376476 4715 scope.go:117] "RemoveContainer" containerID="6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40" Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.376476 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvwbg" Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.399599 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vvwbg"] Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.399707 4715 scope.go:117] "RemoveContainer" containerID="6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40" Dec 04 14:12:11 crc kubenswrapper[4715]: E1204 14:12:11.400784 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40\": container with ID starting with 6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40 not found: ID does not exist" containerID="6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40" Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.400830 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40"} err="failed to get container status \"6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40\": rpc error: code = NotFound desc = could not find container \"6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40\": container with ID starting with 6c923122985f32f2037726a0165a31d8d030ca4cfefa78b2c371615e8a205a40 not found: ID does not exist" Dec 04 14:12:11 crc kubenswrapper[4715]: I1204 14:12:11.401724 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vvwbg"] Dec 04 14:12:12 crc kubenswrapper[4715]: I1204 14:12:12.383347 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerID="145ea078402d171b807e09c9246e462c8bf3b8c8a274a1c20af5dc74a95a41ad" exitCode=0 Dec 04 14:12:12 crc kubenswrapper[4715]: I1204 14:12:12.383616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" event={"ID":"e7c6ddac-1848-4449-8fbe-69d10990b25a","Type":"ContainerDied","Data":"145ea078402d171b807e09c9246e462c8bf3b8c8a274a1c20af5dc74a95a41ad"} Dec 04 14:12:13 crc kubenswrapper[4715]: I1204 14:12:13.189251 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" path="/var/lib/kubelet/pods/c3b08994-cb44-48a9-87d1-af33d2b511b6/volumes" Dec 04 14:12:13 crc kubenswrapper[4715]: I1204 14:12:13.392646 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerID="49fe84adb3fb418a05a976837f4f5cf86da8f88bee7d4ab0827361dad8be5366" exitCode=0 Dec 04 14:12:13 crc kubenswrapper[4715]: I1204 14:12:13.392692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" event={"ID":"e7c6ddac-1848-4449-8fbe-69d10990b25a","Type":"ContainerDied","Data":"49fe84adb3fb418a05a976837f4f5cf86da8f88bee7d4ab0827361dad8be5366"} Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.688451 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.826135 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-bundle\") pod \"e7c6ddac-1848-4449-8fbe-69d10990b25a\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.826215 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-util\") pod \"e7c6ddac-1848-4449-8fbe-69d10990b25a\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.826257 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88dxf\" (UniqueName: \"kubernetes.io/projected/e7c6ddac-1848-4449-8fbe-69d10990b25a-kube-api-access-88dxf\") pod \"e7c6ddac-1848-4449-8fbe-69d10990b25a\" (UID: \"e7c6ddac-1848-4449-8fbe-69d10990b25a\") " Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.827384 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-bundle" (OuterVolumeSpecName: "bundle") pod "e7c6ddac-1848-4449-8fbe-69d10990b25a" (UID: "e7c6ddac-1848-4449-8fbe-69d10990b25a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.832428 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7c6ddac-1848-4449-8fbe-69d10990b25a-kube-api-access-88dxf" (OuterVolumeSpecName: "kube-api-access-88dxf") pod "e7c6ddac-1848-4449-8fbe-69d10990b25a" (UID: "e7c6ddac-1848-4449-8fbe-69d10990b25a"). InnerVolumeSpecName "kube-api-access-88dxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.850568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-util" (OuterVolumeSpecName: "util") pod "e7c6ddac-1848-4449-8fbe-69d10990b25a" (UID: "e7c6ddac-1848-4449-8fbe-69d10990b25a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.927978 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.928290 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88dxf\" (UniqueName: \"kubernetes.io/projected/e7c6ddac-1848-4449-8fbe-69d10990b25a-kube-api-access-88dxf\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4715]: I1204 14:12:14.928370 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7c6ddac-1848-4449-8fbe-69d10990b25a-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:15 crc kubenswrapper[4715]: I1204 14:12:15.414935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" event={"ID":"e7c6ddac-1848-4449-8fbe-69d10990b25a","Type":"ContainerDied","Data":"035a8a4fc708ca75486febc0144c614303df27b3598b5dfe1d76ba8fff71dbb1"} Dec 04 14:12:15 crc kubenswrapper[4715]: I1204 14:12:15.414999 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="035a8a4fc708ca75486febc0144c614303df27b3598b5dfe1d76ba8fff71dbb1" Dec 04 14:12:15 crc kubenswrapper[4715]: I1204 14:12:15.415008 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.552419 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn"] Dec 04 14:12:25 crc kubenswrapper[4715]: E1204 14:12:25.553241 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="util" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553257 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="util" Dec 04 14:12:25 crc kubenswrapper[4715]: E1204 14:12:25.553279 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerName="console" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553287 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerName="console" Dec 04 14:12:25 crc kubenswrapper[4715]: E1204 14:12:25.553298 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="pull" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553306 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="pull" Dec 04 14:12:25 crc kubenswrapper[4715]: E1204 14:12:25.553319 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="extract" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553326 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="extract" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553468 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c6ddac-1848-4449-8fbe-69d10990b25a" containerName="extract" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553478 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b08994-cb44-48a9-87d1-af33d2b511b6" containerName="console" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.553931 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.558198 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9hb9k" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.561726 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.562124 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.562367 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.562711 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.572753 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn"] Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.586755 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46ld7\" (UniqueName: \"kubernetes.io/projected/a57d59fd-4f87-4221-a861-8a4cde8314d6-kube-api-access-46ld7\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.586843 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a57d59fd-4f87-4221-a861-8a4cde8314d6-webhook-cert\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.586874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a57d59fd-4f87-4221-a861-8a4cde8314d6-apiservice-cert\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.687947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46ld7\" (UniqueName: \"kubernetes.io/projected/a57d59fd-4f87-4221-a861-8a4cde8314d6-kube-api-access-46ld7\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.688070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a57d59fd-4f87-4221-a861-8a4cde8314d6-webhook-cert\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.688100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a57d59fd-4f87-4221-a861-8a4cde8314d6-apiservice-cert\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.700213 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a57d59fd-4f87-4221-a861-8a4cde8314d6-webhook-cert\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.700661 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a57d59fd-4f87-4221-a861-8a4cde8314d6-apiservice-cert\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.714994 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46ld7\" (UniqueName: \"kubernetes.io/projected/a57d59fd-4f87-4221-a861-8a4cde8314d6-kube-api-access-46ld7\") pod \"metallb-operator-controller-manager-5d7c69d77f-4r8gn\" (UID: \"a57d59fd-4f87-4221-a861-8a4cde8314d6\") " pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.870504 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.890854 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t"] Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.891755 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.898528 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.898881 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.899210 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pvbsf" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.915411 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t"] Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.992910 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0346d3fa-78df-4769-9c41-b1b1ca9682b9-webhook-cert\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.992994 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0346d3fa-78df-4769-9c41-b1b1ca9682b9-apiservice-cert\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:25 crc kubenswrapper[4715]: I1204 14:12:25.993063 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqzcd\" (UniqueName: \"kubernetes.io/projected/0346d3fa-78df-4769-9c41-b1b1ca9682b9-kube-api-access-nqzcd\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.094618 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0346d3fa-78df-4769-9c41-b1b1ca9682b9-webhook-cert\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.094678 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0346d3fa-78df-4769-9c41-b1b1ca9682b9-apiservice-cert\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.094703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqzcd\" (UniqueName: \"kubernetes.io/projected/0346d3fa-78df-4769-9c41-b1b1ca9682b9-kube-api-access-nqzcd\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.101022 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0346d3fa-78df-4769-9c41-b1b1ca9682b9-apiservice-cert\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.107259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0346d3fa-78df-4769-9c41-b1b1ca9682b9-webhook-cert\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.114694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqzcd\" (UniqueName: \"kubernetes.io/projected/0346d3fa-78df-4769-9c41-b1b1ca9682b9-kube-api-access-nqzcd\") pod \"metallb-operator-webhook-server-78dd5c7fc8-vml6t\" (UID: \"0346d3fa-78df-4769-9c41-b1b1ca9682b9\") " pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.184779 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn"] Dec 04 14:12:26 crc kubenswrapper[4715]: W1204 14:12:26.185750 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda57d59fd_4f87_4221_a861_8a4cde8314d6.slice/crio-15af048d2ce6897b9f8e60dacb3a4bd9a7612adde677ddc4f6425e7ae7f807f3 WatchSource:0}: Error finding container 15af048d2ce6897b9f8e60dacb3a4bd9a7612adde677ddc4f6425e7ae7f807f3: Status 404 returned error can't find the container with id 15af048d2ce6897b9f8e60dacb3a4bd9a7612adde677ddc4f6425e7ae7f807f3 Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.258470 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.476942 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" event={"ID":"a57d59fd-4f87-4221-a861-8a4cde8314d6","Type":"ContainerStarted","Data":"15af048d2ce6897b9f8e60dacb3a4bd9a7612adde677ddc4f6425e7ae7f807f3"} Dec 04 14:12:26 crc kubenswrapper[4715]: I1204 14:12:26.519170 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t"] Dec 04 14:12:26 crc kubenswrapper[4715]: W1204 14:12:26.527254 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0346d3fa_78df_4769_9c41_b1b1ca9682b9.slice/crio-1985c5e6c2c9216dc593cc6ac7bb6ff1df57019817bd7552e85c51cda4dce060 WatchSource:0}: Error finding container 1985c5e6c2c9216dc593cc6ac7bb6ff1df57019817bd7552e85c51cda4dce060: Status 404 returned error can't find the container with id 1985c5e6c2c9216dc593cc6ac7bb6ff1df57019817bd7552e85c51cda4dce060 Dec 04 14:12:27 crc kubenswrapper[4715]: I1204 14:12:27.483060 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" event={"ID":"0346d3fa-78df-4769-9c41-b1b1ca9682b9","Type":"ContainerStarted","Data":"1985c5e6c2c9216dc593cc6ac7bb6ff1df57019817bd7552e85c51cda4dce060"} Dec 04 14:12:29 crc kubenswrapper[4715]: I1204 14:12:29.513919 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" event={"ID":"a57d59fd-4f87-4221-a861-8a4cde8314d6","Type":"ContainerStarted","Data":"a2fb505587c845da7d1fd9e0dd7524d76642ff4e5fcbaf73ffe5d16f536db89d"} Dec 04 14:12:29 crc kubenswrapper[4715]: I1204 14:12:29.514454 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:12:29 crc kubenswrapper[4715]: I1204 14:12:29.558788 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" podStartSLOduration=1.8884620220000001 podStartE2EDuration="4.558759982s" podCreationTimestamp="2025-12-04 14:12:25 +0000 UTC" firstStartedPulling="2025-12-04 14:12:26.187870673 +0000 UTC m=+903.256588888" lastFinishedPulling="2025-12-04 14:12:28.858168623 +0000 UTC m=+905.926886848" observedRunningTime="2025-12-04 14:12:29.557435627 +0000 UTC m=+906.626153862" watchObservedRunningTime="2025-12-04 14:12:29.558759982 +0000 UTC m=+906.627478207" Dec 04 14:12:35 crc kubenswrapper[4715]: I1204 14:12:35.558750 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" event={"ID":"0346d3fa-78df-4769-9c41-b1b1ca9682b9","Type":"ContainerStarted","Data":"fdb82b515453cbd25e5408ecc11dac63fb94b98d715f81436b973c1d8130708d"} Dec 04 14:12:35 crc kubenswrapper[4715]: I1204 14:12:35.559178 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:12:35 crc kubenswrapper[4715]: I1204 14:12:35.584997 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" podStartSLOduration=1.91645808 podStartE2EDuration="10.584968081s" podCreationTimestamp="2025-12-04 14:12:25 +0000 UTC" firstStartedPulling="2025-12-04 14:12:26.530654307 +0000 UTC m=+903.599372522" lastFinishedPulling="2025-12-04 14:12:35.199164308 +0000 UTC m=+912.267882523" observedRunningTime="2025-12-04 14:12:35.579778153 +0000 UTC m=+912.648496378" watchObservedRunningTime="2025-12-04 14:12:35.584968081 +0000 UTC m=+912.653686296" Dec 04 14:12:46 crc kubenswrapper[4715]: I1204 14:12:46.263342 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-78dd5c7fc8-vml6t" Dec 04 14:13:05 crc kubenswrapper[4715]: I1204 14:13:05.874142 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d7c69d77f-4r8gn" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.673761 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb"] Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.675244 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.684768 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-vmq8w" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.684942 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.698920 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6hp55"] Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.701865 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.702572 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb"] Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.703913 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.704210 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742252 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-sockets\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742354 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-startup\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742392 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d254ca37-56e1-470f-839b-cc07aa846cd3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-znkxb\" (UID: \"d254ca37-56e1-470f-839b-cc07aa846cd3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742428 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-reloader\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzl96\" (UniqueName: \"kubernetes.io/projected/a4126a74-57d7-4987-bd7a-091b5e6f3806-kube-api-access-qzl96\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742471 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics-certs\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742496 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9st72\" (UniqueName: \"kubernetes.io/projected/d254ca37-56e1-470f-839b-cc07aa846cd3-kube-api-access-9st72\") pod \"frr-k8s-webhook-server-7fcb986d4-znkxb\" (UID: \"d254ca37-56e1-470f-839b-cc07aa846cd3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.742533 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-conf\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.773228 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mx2g2"] Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.774371 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.776140 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.779387 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.779512 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.779674 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-4vgw4" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.785949 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-kp8gj"] Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.786970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.788676 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.790392 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-kp8gj"] Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843619 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1de7de85-bcce-4132-9885-45d7996f6845-metallb-excludel2\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843688 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-conf\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843724 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4a52cd8-e319-4888-8920-a69547290700-metrics-certs\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843758 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843790 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4a52cd8-e319-4888-8920-a69547290700-cert\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-metrics-certs\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843861 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-sockets\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843890 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-startup\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843915 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhfhf\" (UniqueName: \"kubernetes.io/projected/1de7de85-bcce-4132-9885-45d7996f6845-kube-api-access-hhfhf\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843947 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lhqd\" (UniqueName: \"kubernetes.io/projected/c4a52cd8-e319-4888-8920-a69547290700-kube-api-access-7lhqd\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843973 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.843996 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d254ca37-56e1-470f-839b-cc07aa846cd3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-znkxb\" (UID: \"d254ca37-56e1-470f-839b-cc07aa846cd3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844050 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-reloader\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics-certs\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzl96\" (UniqueName: \"kubernetes.io/projected/a4126a74-57d7-4987-bd7a-091b5e6f3806-kube-api-access-qzl96\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844114 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9st72\" (UniqueName: \"kubernetes.io/projected/d254ca37-56e1-470f-839b-cc07aa846cd3-kube-api-access-9st72\") pod \"frr-k8s-webhook-server-7fcb986d4-znkxb\" (UID: \"d254ca37-56e1-470f-839b-cc07aa846cd3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844173 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-conf\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844237 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-sockets\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: E1204 14:13:06.844520 4715 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 04 14:13:06 crc kubenswrapper[4715]: E1204 14:13:06.844566 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics-certs podName:a4126a74-57d7-4987-bd7a-091b5e6f3806 nodeName:}" failed. No retries permitted until 2025-12-04 14:13:07.344549335 +0000 UTC m=+944.413267610 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics-certs") pod "frr-k8s-6hp55" (UID: "a4126a74-57d7-4987-bd7a-091b5e6f3806") : secret "frr-k8s-certs-secret" not found Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.844584 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a4126a74-57d7-4987-bd7a-091b5e6f3806-reloader\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.845156 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a4126a74-57d7-4987-bd7a-091b5e6f3806-frr-startup\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.859105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d254ca37-56e1-470f-839b-cc07aa846cd3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-znkxb\" (UID: \"d254ca37-56e1-470f-839b-cc07aa846cd3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.861483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzl96\" (UniqueName: \"kubernetes.io/projected/a4126a74-57d7-4987-bd7a-091b5e6f3806-kube-api-access-qzl96\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.865869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9st72\" (UniqueName: \"kubernetes.io/projected/d254ca37-56e1-470f-839b-cc07aa846cd3-kube-api-access-9st72\") pod \"frr-k8s-webhook-server-7fcb986d4-znkxb\" (UID: \"d254ca37-56e1-470f-839b-cc07aa846cd3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.944835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4a52cd8-e319-4888-8920-a69547290700-metrics-certs\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.944886 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4a52cd8-e319-4888-8920-a69547290700-cert\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.944905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-metrics-certs\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.944940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhfhf\" (UniqueName: \"kubernetes.io/projected/1de7de85-bcce-4132-9885-45d7996f6845-kube-api-access-hhfhf\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.944962 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lhqd\" (UniqueName: \"kubernetes.io/projected/c4a52cd8-e319-4888-8920-a69547290700-kube-api-access-7lhqd\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.944984 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.945053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1de7de85-bcce-4132-9885-45d7996f6845-metallb-excludel2\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.945698 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1de7de85-bcce-4132-9885-45d7996f6845-metallb-excludel2\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: E1204 14:13:06.946316 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 14:13:06 crc kubenswrapper[4715]: E1204 14:13:06.946358 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist podName:1de7de85-bcce-4132-9885-45d7996f6845 nodeName:}" failed. No retries permitted until 2025-12-04 14:13:07.446345214 +0000 UTC m=+944.515063429 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist") pod "speaker-mx2g2" (UID: "1de7de85-bcce-4132-9885-45d7996f6845") : secret "metallb-memberlist" not found Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.953292 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4a52cd8-e319-4888-8920-a69547290700-metrics-certs\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.954446 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4a52cd8-e319-4888-8920-a69547290700-cert\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.954771 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-metrics-certs\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.962322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lhqd\" (UniqueName: \"kubernetes.io/projected/c4a52cd8-e319-4888-8920-a69547290700-kube-api-access-7lhqd\") pod \"controller-f8648f98b-kp8gj\" (UID: \"c4a52cd8-e319-4888-8920-a69547290700\") " pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:06 crc kubenswrapper[4715]: I1204 14:13:06.963288 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhfhf\" (UniqueName: \"kubernetes.io/projected/1de7de85-bcce-4132-9885-45d7996f6845-kube-api-access-hhfhf\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.011675 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.129490 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.350224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics-certs\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.354088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4126a74-57d7-4987-bd7a-091b5e6f3806-metrics-certs\") pod \"frr-k8s-6hp55\" (UID: \"a4126a74-57d7-4987-bd7a-091b5e6f3806\") " pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.452829 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:07 crc kubenswrapper[4715]: E1204 14:13:07.453085 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 14:13:07 crc kubenswrapper[4715]: E1204 14:13:07.453412 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist podName:1de7de85-bcce-4132-9885-45d7996f6845 nodeName:}" failed. No retries permitted until 2025-12-04 14:13:08.453341053 +0000 UTC m=+945.522059268 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist") pod "speaker-mx2g2" (UID: "1de7de85-bcce-4132-9885-45d7996f6845") : secret "metallb-memberlist" not found Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.534858 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-kp8gj"] Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.590201 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb"] Dec 04 14:13:07 crc kubenswrapper[4715]: W1204 14:13:07.601506 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd254ca37_56e1_470f_839b_cc07aa846cd3.slice/crio-e6e73a406f491c54056522a3c7007773c81d8876dae1c85954f2f5d46d4b13cf WatchSource:0}: Error finding container e6e73a406f491c54056522a3c7007773c81d8876dae1c85954f2f5d46d4b13cf: Status 404 returned error can't find the container with id e6e73a406f491c54056522a3c7007773c81d8876dae1c85954f2f5d46d4b13cf Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.619323 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.865059 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" event={"ID":"d254ca37-56e1-470f-839b-cc07aa846cd3","Type":"ContainerStarted","Data":"e6e73a406f491c54056522a3c7007773c81d8876dae1c85954f2f5d46d4b13cf"} Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.866664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-kp8gj" event={"ID":"c4a52cd8-e319-4888-8920-a69547290700","Type":"ContainerStarted","Data":"2382b24fd6861e6dc469d4c1da52296833593f69b1955b2b7450d3c0c1dbc70d"} Dec 04 14:13:07 crc kubenswrapper[4715]: I1204 14:13:07.866707 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-kp8gj" event={"ID":"c4a52cd8-e319-4888-8920-a69547290700","Type":"ContainerStarted","Data":"6e7edf47b013302f5434a1edd6c76fba5efe2da845bbdcecc5b883ecb4824a59"} Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.467904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.473435 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1de7de85-bcce-4132-9885-45d7996f6845-memberlist\") pod \"speaker-mx2g2\" (UID: \"1de7de85-bcce-4132-9885-45d7996f6845\") " pod="metallb-system/speaker-mx2g2" Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.589564 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mx2g2" Dec 04 14:13:08 crc kubenswrapper[4715]: W1204 14:13:08.615307 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1de7de85_bcce_4132_9885_45d7996f6845.slice/crio-31929c58b7308f9e92f08028c422ae688808bd61efcb3d0424e0ede962c3cd0f WatchSource:0}: Error finding container 31929c58b7308f9e92f08028c422ae688808bd61efcb3d0424e0ede962c3cd0f: Status 404 returned error can't find the container with id 31929c58b7308f9e92f08028c422ae688808bd61efcb3d0424e0ede962c3cd0f Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.877738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"c6a9d515be75e1ad40aedfa6acbaa10f5899fae40609de933744cfea30e7da41"} Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.879440 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-kp8gj" event={"ID":"c4a52cd8-e319-4888-8920-a69547290700","Type":"ContainerStarted","Data":"41b38712d135846aa3186adbe849f35c88a08ad4e5008edec5e587c0d2daa7a1"} Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.880375 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.889447 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mx2g2" event={"ID":"1de7de85-bcce-4132-9885-45d7996f6845","Type":"ContainerStarted","Data":"31929c58b7308f9e92f08028c422ae688808bd61efcb3d0424e0ede962c3cd0f"} Dec 04 14:13:08 crc kubenswrapper[4715]: I1204 14:13:08.908486 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-kp8gj" podStartSLOduration=2.908462043 podStartE2EDuration="2.908462043s" podCreationTimestamp="2025-12-04 14:13:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:08.902985656 +0000 UTC m=+945.971703871" watchObservedRunningTime="2025-12-04 14:13:08.908462043 +0000 UTC m=+945.977180258" Dec 04 14:13:10 crc kubenswrapper[4715]: I1204 14:13:10.289188 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mx2g2" event={"ID":"1de7de85-bcce-4132-9885-45d7996f6845","Type":"ContainerStarted","Data":"d003e3f76f01818f1624bf884f6a108fe32c44abd5c649bf90b9150f478376fa"} Dec 04 14:13:10 crc kubenswrapper[4715]: I1204 14:13:10.289522 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mx2g2" Dec 04 14:13:10 crc kubenswrapper[4715]: I1204 14:13:10.289540 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mx2g2" event={"ID":"1de7de85-bcce-4132-9885-45d7996f6845","Type":"ContainerStarted","Data":"c80f278fff1d02598839df94d5103e24425fa669379ca3c7d5e4bcd73b881e67"} Dec 04 14:13:10 crc kubenswrapper[4715]: I1204 14:13:10.309817 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mx2g2" podStartSLOduration=4.309798064 podStartE2EDuration="4.309798064s" podCreationTimestamp="2025-12-04 14:13:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:10.305728006 +0000 UTC m=+947.374446231" watchObservedRunningTime="2025-12-04 14:13:10.309798064 +0000 UTC m=+947.378516279" Dec 04 14:13:17 crc kubenswrapper[4715]: I1204 14:13:17.135509 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-kp8gj" Dec 04 14:13:18 crc kubenswrapper[4715]: I1204 14:13:18.595730 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mx2g2" Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.931565 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rpzk2"] Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.932768 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.942325 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.942577 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.942714 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-skjfs" Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.945931 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rpzk2"] Dec 04 14:13:21 crc kubenswrapper[4715]: I1204 14:13:21.954061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcl4f\" (UniqueName: \"kubernetes.io/projected/4915974a-a9a1-4f30-bf50-b9d0a64d385d-kube-api-access-rcl4f\") pod \"openstack-operator-index-rpzk2\" (UID: \"4915974a-a9a1-4f30-bf50-b9d0a64d385d\") " pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.054828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcl4f\" (UniqueName: \"kubernetes.io/projected/4915974a-a9a1-4f30-bf50-b9d0a64d385d-kube-api-access-rcl4f\") pod \"openstack-operator-index-rpzk2\" (UID: \"4915974a-a9a1-4f30-bf50-b9d0a64d385d\") " pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.080515 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcl4f\" (UniqueName: \"kubernetes.io/projected/4915974a-a9a1-4f30-bf50-b9d0a64d385d-kube-api-access-rcl4f\") pod \"openstack-operator-index-rpzk2\" (UID: \"4915974a-a9a1-4f30-bf50-b9d0a64d385d\") " pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.301974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.739649 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" event={"ID":"d254ca37-56e1-470f-839b-cc07aa846cd3","Type":"ContainerStarted","Data":"adf9e9fe36398f843db2163272309003376e7e0324338f26d00e9badc2c744e7"} Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.740025 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.742793 4715 generic.go:334] "Generic (PLEG): container finished" podID="a4126a74-57d7-4987-bd7a-091b5e6f3806" containerID="d4849e995ef633451c533535caac89d14fd502634b862336dd3ed24a03339801" exitCode=0 Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.742848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerDied","Data":"d4849e995ef633451c533535caac89d14fd502634b862336dd3ed24a03339801"} Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.760520 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" podStartSLOduration=2.507384232 podStartE2EDuration="16.76050142s" podCreationTimestamp="2025-12-04 14:13:06 +0000 UTC" firstStartedPulling="2025-12-04 14:13:07.604300305 +0000 UTC m=+944.673018520" lastFinishedPulling="2025-12-04 14:13:21.857417493 +0000 UTC m=+958.926135708" observedRunningTime="2025-12-04 14:13:22.755969409 +0000 UTC m=+959.824687624" watchObservedRunningTime="2025-12-04 14:13:22.76050142 +0000 UTC m=+959.829219635" Dec 04 14:13:22 crc kubenswrapper[4715]: I1204 14:13:22.900354 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rpzk2"] Dec 04 14:13:23 crc kubenswrapper[4715]: I1204 14:13:23.777098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rpzk2" event={"ID":"4915974a-a9a1-4f30-bf50-b9d0a64d385d","Type":"ContainerStarted","Data":"e23bcc5e1df6705f2558cbf71465a06e4cb04dd88ce42d253faed48c5384f5af"} Dec 04 14:13:24 crc kubenswrapper[4715]: I1204 14:13:24.786328 4715 generic.go:334] "Generic (PLEG): container finished" podID="a4126a74-57d7-4987-bd7a-091b5e6f3806" containerID="4da9ee5df7a9122075ffecde28a596b4e9bfcb441ed05242ecf57b79d766e4d0" exitCode=0 Dec 04 14:13:24 crc kubenswrapper[4715]: I1204 14:13:24.786374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerDied","Data":"4da9ee5df7a9122075ffecde28a596b4e9bfcb441ed05242ecf57b79d766e4d0"} Dec 04 14:13:25 crc kubenswrapper[4715]: I1204 14:13:25.310691 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rpzk2"] Dec 04 14:13:25 crc kubenswrapper[4715]: I1204 14:13:25.923910 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pnqdc"] Dec 04 14:13:25 crc kubenswrapper[4715]: I1204 14:13:25.924981 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:25 crc kubenswrapper[4715]: I1204 14:13:25.929001 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pnqdc"] Dec 04 14:13:26 crc kubenswrapper[4715]: I1204 14:13:26.116397 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n44b\" (UniqueName: \"kubernetes.io/projected/f7e73b69-ec29-4bdb-9374-25916f8df390-kube-api-access-4n44b\") pod \"openstack-operator-index-pnqdc\" (UID: \"f7e73b69-ec29-4bdb-9374-25916f8df390\") " pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:26 crc kubenswrapper[4715]: I1204 14:13:26.217922 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n44b\" (UniqueName: \"kubernetes.io/projected/f7e73b69-ec29-4bdb-9374-25916f8df390-kube-api-access-4n44b\") pod \"openstack-operator-index-pnqdc\" (UID: \"f7e73b69-ec29-4bdb-9374-25916f8df390\") " pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:26 crc kubenswrapper[4715]: I1204 14:13:26.236611 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n44b\" (UniqueName: \"kubernetes.io/projected/f7e73b69-ec29-4bdb-9374-25916f8df390-kube-api-access-4n44b\") pod \"openstack-operator-index-pnqdc\" (UID: \"f7e73b69-ec29-4bdb-9374-25916f8df390\") " pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:26 crc kubenswrapper[4715]: I1204 14:13:26.245278 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:26 crc kubenswrapper[4715]: I1204 14:13:26.799949 4715 generic.go:334] "Generic (PLEG): container finished" podID="a4126a74-57d7-4987-bd7a-091b5e6f3806" containerID="f88b2b60536bcec119e4dc778bc9e10ae651ab3e968f3e9b4d2832c931cf1f48" exitCode=0 Dec 04 14:13:26 crc kubenswrapper[4715]: I1204 14:13:26.800163 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerDied","Data":"f88b2b60536bcec119e4dc778bc9e10ae651ab3e968f3e9b4d2832c931cf1f48"} Dec 04 14:13:27 crc kubenswrapper[4715]: I1204 14:13:27.649984 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pnqdc"] Dec 04 14:13:27 crc kubenswrapper[4715]: I1204 14:13:27.829110 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pnqdc" event={"ID":"f7e73b69-ec29-4bdb-9374-25916f8df390","Type":"ContainerStarted","Data":"6c034fb28708d873fcd0a7ee947866217f44bce5919effe2a21945cb865d2752"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.839097 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"3515d10acf68bd0f33aa451606b3948433858f9ac27909baff245084bfcb4147"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.839416 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"a136bbd5004f503e7e65b73da939163bb714995a09eaf06dae08550d7caebd03"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.839427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"84cf5a1673d6b4690b9fd949033053bf6166766e5470cd079c27581f584fbd58"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.839436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"f0f95e3947f034034bda00791a1f17e68afd61dc65e29f88401a2582bbc1f21a"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.839444 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"81ca7f71b0370318efa47751f2c533efc61c84ef18355be8d0ff5f8b2ee77c39"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.840293 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rpzk2" event={"ID":"4915974a-a9a1-4f30-bf50-b9d0a64d385d","Type":"ContainerStarted","Data":"21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.840383 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-rpzk2" podUID="4915974a-a9a1-4f30-bf50-b9d0a64d385d" containerName="registry-server" containerID="cri-o://21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7" gracePeriod=2 Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.841789 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pnqdc" event={"ID":"f7e73b69-ec29-4bdb-9374-25916f8df390","Type":"ContainerStarted","Data":"3d6cd6c25fd9e72708702141518801a816b564332f91c0b1d133cdfc021108f7"} Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.875935 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rpzk2" podStartSLOduration=2.922688215 podStartE2EDuration="7.875917121s" podCreationTimestamp="2025-12-04 14:13:21 +0000 UTC" firstStartedPulling="2025-12-04 14:13:22.911140872 +0000 UTC m=+959.979859087" lastFinishedPulling="2025-12-04 14:13:27.864369778 +0000 UTC m=+964.933087993" observedRunningTime="2025-12-04 14:13:28.857401987 +0000 UTC m=+965.926120212" watchObservedRunningTime="2025-12-04 14:13:28.875917121 +0000 UTC m=+965.944635336" Dec 04 14:13:28 crc kubenswrapper[4715]: I1204 14:13:28.877685 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pnqdc" podStartSLOduration=3.797676091 podStartE2EDuration="3.877677028s" podCreationTimestamp="2025-12-04 14:13:25 +0000 UTC" firstStartedPulling="2025-12-04 14:13:27.809837252 +0000 UTC m=+964.878555467" lastFinishedPulling="2025-12-04 14:13:27.889838189 +0000 UTC m=+964.958556404" observedRunningTime="2025-12-04 14:13:28.874422301 +0000 UTC m=+965.943140516" watchObservedRunningTime="2025-12-04 14:13:28.877677028 +0000 UTC m=+965.946395243" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.420810 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.604499 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcl4f\" (UniqueName: \"kubernetes.io/projected/4915974a-a9a1-4f30-bf50-b9d0a64d385d-kube-api-access-rcl4f\") pod \"4915974a-a9a1-4f30-bf50-b9d0a64d385d\" (UID: \"4915974a-a9a1-4f30-bf50-b9d0a64d385d\") " Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.611363 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4915974a-a9a1-4f30-bf50-b9d0a64d385d-kube-api-access-rcl4f" (OuterVolumeSpecName: "kube-api-access-rcl4f") pod "4915974a-a9a1-4f30-bf50-b9d0a64d385d" (UID: "4915974a-a9a1-4f30-bf50-b9d0a64d385d"). InnerVolumeSpecName "kube-api-access-rcl4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.706003 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcl4f\" (UniqueName: \"kubernetes.io/projected/4915974a-a9a1-4f30-bf50-b9d0a64d385d-kube-api-access-rcl4f\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.854851 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6hp55" event={"ID":"a4126a74-57d7-4987-bd7a-091b5e6f3806","Type":"ContainerStarted","Data":"de06640e3305ea74358ae59fca891295f8eb4fad53a3e195ca705579a6e6f145"} Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.855262 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.857479 4715 generic.go:334] "Generic (PLEG): container finished" podID="4915974a-a9a1-4f30-bf50-b9d0a64d385d" containerID="21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7" exitCode=0 Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.857607 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rpzk2" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.857582 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rpzk2" event={"ID":"4915974a-a9a1-4f30-bf50-b9d0a64d385d","Type":"ContainerDied","Data":"21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7"} Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.857838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rpzk2" event={"ID":"4915974a-a9a1-4f30-bf50-b9d0a64d385d","Type":"ContainerDied","Data":"e23bcc5e1df6705f2558cbf71465a06e4cb04dd88ce42d253faed48c5384f5af"} Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.857925 4715 scope.go:117] "RemoveContainer" containerID="21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.888377 4715 scope.go:117] "RemoveContainer" containerID="21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.889562 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6hp55" podStartSLOduration=10.174790779 podStartE2EDuration="23.88954935s" podCreationTimestamp="2025-12-04 14:13:06 +0000 UTC" firstStartedPulling="2025-12-04 14:13:08.157675693 +0000 UTC m=+945.226393908" lastFinishedPulling="2025-12-04 14:13:21.872434264 +0000 UTC m=+958.941152479" observedRunningTime="2025-12-04 14:13:29.882895482 +0000 UTC m=+966.951613697" watchObservedRunningTime="2025-12-04 14:13:29.88954935 +0000 UTC m=+966.958267565" Dec 04 14:13:29 crc kubenswrapper[4715]: E1204 14:13:29.890686 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7\": container with ID starting with 21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7 not found: ID does not exist" containerID="21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.890764 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7"} err="failed to get container status \"21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7\": rpc error: code = NotFound desc = could not find container \"21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7\": container with ID starting with 21077e27fe57064e796b041502ed63fb96e09b692f7a866681acf2077cdacdf7 not found: ID does not exist" Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.899743 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rpzk2"] Dec 04 14:13:29 crc kubenswrapper[4715]: I1204 14:13:29.903658 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-rpzk2"] Dec 04 14:13:31 crc kubenswrapper[4715]: I1204 14:13:31.187925 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4915974a-a9a1-4f30-bf50-b9d0a64d385d" path="/var/lib/kubelet/pods/4915974a-a9a1-4f30-bf50-b9d0a64d385d/volumes" Dec 04 14:13:32 crc kubenswrapper[4715]: I1204 14:13:32.619869 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:32 crc kubenswrapper[4715]: I1204 14:13:32.660669 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:36 crc kubenswrapper[4715]: I1204 14:13:36.246286 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:36 crc kubenswrapper[4715]: I1204 14:13:36.246594 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:36 crc kubenswrapper[4715]: I1204 14:13:36.273310 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:36 crc kubenswrapper[4715]: I1204 14:13:36.926276 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-pnqdc" Dec 04 14:13:37 crc kubenswrapper[4715]: I1204 14:13:37.016207 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znkxb" Dec 04 14:13:37 crc kubenswrapper[4715]: I1204 14:13:37.623159 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6hp55" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.249909 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw"] Dec 04 14:13:43 crc kubenswrapper[4715]: E1204 14:13:43.250467 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4915974a-a9a1-4f30-bf50-b9d0a64d385d" containerName="registry-server" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.250483 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4915974a-a9a1-4f30-bf50-b9d0a64d385d" containerName="registry-server" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.250644 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4915974a-a9a1-4f30-bf50-b9d0a64d385d" containerName="registry-server" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.254356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.257991 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw"] Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.258536 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cw9dp" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.392751 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-util\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.392802 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qdxv\" (UniqueName: \"kubernetes.io/projected/1d09748d-bd97-4abd-ade6-92e80c61ffbd-kube-api-access-2qdxv\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.392854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-bundle\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.494575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-bundle\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.494684 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-util\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.494708 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qdxv\" (UniqueName: \"kubernetes.io/projected/1d09748d-bd97-4abd-ade6-92e80c61ffbd-kube-api-access-2qdxv\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.495156 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-bundle\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.495321 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-util\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.514006 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qdxv\" (UniqueName: \"kubernetes.io/projected/1d09748d-bd97-4abd-ade6-92e80c61ffbd-kube-api-access-2qdxv\") pod \"0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:43 crc kubenswrapper[4715]: I1204 14:13:43.574745 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:44 crc kubenswrapper[4715]: I1204 14:13:44.030734 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw"] Dec 04 14:13:44 crc kubenswrapper[4715]: W1204 14:13:44.039067 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d09748d_bd97_4abd_ade6_92e80c61ffbd.slice/crio-ff9682db91ac45d726fa35cf144d0ac7d0293611a8a57cf07cd6a1c752a53c46 WatchSource:0}: Error finding container ff9682db91ac45d726fa35cf144d0ac7d0293611a8a57cf07cd6a1c752a53c46: Status 404 returned error can't find the container with id ff9682db91ac45d726fa35cf144d0ac7d0293611a8a57cf07cd6a1c752a53c46 Dec 04 14:13:44 crc kubenswrapper[4715]: I1204 14:13:44.947475 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerID="e4f3ccb52feb8c9fb852ea31c589786d6273a4908ee3dd0168a055d95b34500f" exitCode=0 Dec 04 14:13:44 crc kubenswrapper[4715]: I1204 14:13:44.947605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" event={"ID":"1d09748d-bd97-4abd-ade6-92e80c61ffbd","Type":"ContainerDied","Data":"e4f3ccb52feb8c9fb852ea31c589786d6273a4908ee3dd0168a055d95b34500f"} Dec 04 14:13:44 crc kubenswrapper[4715]: I1204 14:13:44.947817 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" event={"ID":"1d09748d-bd97-4abd-ade6-92e80c61ffbd","Type":"ContainerStarted","Data":"ff9682db91ac45d726fa35cf144d0ac7d0293611a8a57cf07cd6a1c752a53c46"} Dec 04 14:13:45 crc kubenswrapper[4715]: I1204 14:13:45.955853 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerID="d88480a7c855fda41f5dc3d744a2792ba0a4ffaeba33c4c09e0a090c24c32521" exitCode=0 Dec 04 14:13:45 crc kubenswrapper[4715]: I1204 14:13:45.955905 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" event={"ID":"1d09748d-bd97-4abd-ade6-92e80c61ffbd","Type":"ContainerDied","Data":"d88480a7c855fda41f5dc3d744a2792ba0a4ffaeba33c4c09e0a090c24c32521"} Dec 04 14:13:46 crc kubenswrapper[4715]: I1204 14:13:46.967503 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerID="2d58dab53942a1870c68155357f9ef7bfa9ae5873648c1e942264e3fefe4eb05" exitCode=0 Dec 04 14:13:46 crc kubenswrapper[4715]: I1204 14:13:46.967578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" event={"ID":"1d09748d-bd97-4abd-ade6-92e80c61ffbd","Type":"ContainerDied","Data":"2d58dab53942a1870c68155357f9ef7bfa9ae5873648c1e942264e3fefe4eb05"} Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.245222 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.369881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qdxv\" (UniqueName: \"kubernetes.io/projected/1d09748d-bd97-4abd-ade6-92e80c61ffbd-kube-api-access-2qdxv\") pod \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.369970 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-util\") pod \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.370068 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-bundle\") pod \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\" (UID: \"1d09748d-bd97-4abd-ade6-92e80c61ffbd\") " Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.371102 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-bundle" (OuterVolumeSpecName: "bundle") pod "1d09748d-bd97-4abd-ade6-92e80c61ffbd" (UID: "1d09748d-bd97-4abd-ade6-92e80c61ffbd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.376858 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d09748d-bd97-4abd-ade6-92e80c61ffbd-kube-api-access-2qdxv" (OuterVolumeSpecName: "kube-api-access-2qdxv") pod "1d09748d-bd97-4abd-ade6-92e80c61ffbd" (UID: "1d09748d-bd97-4abd-ade6-92e80c61ffbd"). InnerVolumeSpecName "kube-api-access-2qdxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.385094 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-util" (OuterVolumeSpecName: "util") pod "1d09748d-bd97-4abd-ade6-92e80c61ffbd" (UID: "1d09748d-bd97-4abd-ade6-92e80c61ffbd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.471302 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.471337 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1d09748d-bd97-4abd-ade6-92e80c61ffbd-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.471346 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qdxv\" (UniqueName: \"kubernetes.io/projected/1d09748d-bd97-4abd-ade6-92e80c61ffbd-kube-api-access-2qdxv\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.984230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" event={"ID":"1d09748d-bd97-4abd-ade6-92e80c61ffbd","Type":"ContainerDied","Data":"ff9682db91ac45d726fa35cf144d0ac7d0293611a8a57cf07cd6a1c752a53c46"} Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.984279 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff9682db91ac45d726fa35cf144d0ac7d0293611a8a57cf07cd6a1c752a53c46" Dec 04 14:13:48 crc kubenswrapper[4715]: I1204 14:13:48.984283 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.451635 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-52hfz"] Dec 04 14:13:49 crc kubenswrapper[4715]: E1204 14:13:49.452238 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="pull" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.452254 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="pull" Dec 04 14:13:49 crc kubenswrapper[4715]: E1204 14:13:49.452269 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="util" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.452277 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="util" Dec 04 14:13:49 crc kubenswrapper[4715]: E1204 14:13:49.452291 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="extract" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.452298 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="extract" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.452435 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d09748d-bd97-4abd-ade6-92e80c61ffbd" containerName="extract" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.453464 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.468117 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-52hfz"] Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.487500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdrsc\" (UniqueName: \"kubernetes.io/projected/6b2f3a01-4232-4457-8034-c18cad1517cf-kube-api-access-hdrsc\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.487608 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-utilities\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.487655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-catalog-content\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.588839 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-utilities\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.588917 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-catalog-content\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.588975 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdrsc\" (UniqueName: \"kubernetes.io/projected/6b2f3a01-4232-4457-8034-c18cad1517cf-kube-api-access-hdrsc\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.589796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-utilities\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.589858 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-catalog-content\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.612121 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdrsc\" (UniqueName: \"kubernetes.io/projected/6b2f3a01-4232-4457-8034-c18cad1517cf-kube-api-access-hdrsc\") pod \"community-operators-52hfz\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:49 crc kubenswrapper[4715]: I1204 14:13:49.779415 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:50 crc kubenswrapper[4715]: I1204 14:13:50.369890 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-52hfz"] Dec 04 14:13:51 crc kubenswrapper[4715]: I1204 14:13:51.042022 4715 generic.go:334] "Generic (PLEG): container finished" podID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerID="4363dbe6fcd80ba00ed73da67c1fbfaf3201e0ec5ebd312a1218fc3dd4fad012" exitCode=0 Dec 04 14:13:51 crc kubenswrapper[4715]: I1204 14:13:51.042158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerDied","Data":"4363dbe6fcd80ba00ed73da67c1fbfaf3201e0ec5ebd312a1218fc3dd4fad012"} Dec 04 14:13:51 crc kubenswrapper[4715]: I1204 14:13:51.042412 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerStarted","Data":"578490950a822360f0079b8e301234bb4b1d05612ac2159a24cb3b56cf8261a5"} Dec 04 14:13:52 crc kubenswrapper[4715]: I1204 14:13:52.049309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerStarted","Data":"c5ede22eeefd8f833e476aab4304a13962c21cf4d2b9630c16014c680f3679de"} Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.058088 4715 generic.go:334] "Generic (PLEG): container finished" podID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerID="c5ede22eeefd8f833e476aab4304a13962c21cf4d2b9630c16014c680f3679de" exitCode=0 Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.058174 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerDied","Data":"c5ede22eeefd8f833e476aab4304a13962c21cf4d2b9630c16014c680f3679de"} Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.237313 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hzf2f"] Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.241647 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.244957 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzf2f"] Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.294692 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-utilities\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.294752 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hptzt\" (UniqueName: \"kubernetes.io/projected/cbf18951-ee68-4717-93c0-ec9a1826e37c-kube-api-access-hptzt\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.294828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-catalog-content\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.396109 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-utilities\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.396360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hptzt\" (UniqueName: \"kubernetes.io/projected/cbf18951-ee68-4717-93c0-ec9a1826e37c-kube-api-access-hptzt\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.396501 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-catalog-content\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.396758 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-utilities\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.396980 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-catalog-content\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.437138 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hptzt\" (UniqueName: \"kubernetes.io/projected/cbf18951-ee68-4717-93c0-ec9a1826e37c-kube-api-access-hptzt\") pod \"redhat-marketplace-hzf2f\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:53 crc kubenswrapper[4715]: I1204 14:13:53.564360 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.174579 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzf2f"] Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.563655 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj"] Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.564913 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.568596 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-t9lnz" Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.589136 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj"] Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.716789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4wl\" (UniqueName: \"kubernetes.io/projected/21b74fe1-833a-4744-b868-6b09c3e93458-kube-api-access-6l4wl\") pod \"openstack-operator-controller-operator-6cf7bc7f5d-kzpwj\" (UID: \"21b74fe1-833a-4744-b868-6b09c3e93458\") " pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.818515 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4wl\" (UniqueName: \"kubernetes.io/projected/21b74fe1-833a-4744-b868-6b09c3e93458-kube-api-access-6l4wl\") pod \"openstack-operator-controller-operator-6cf7bc7f5d-kzpwj\" (UID: \"21b74fe1-833a-4744-b868-6b09c3e93458\") " pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.853464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4wl\" (UniqueName: \"kubernetes.io/projected/21b74fe1-833a-4744-b868-6b09c3e93458-kube-api-access-6l4wl\") pod \"openstack-operator-controller-operator-6cf7bc7f5d-kzpwj\" (UID: \"21b74fe1-833a-4744-b868-6b09c3e93458\") " pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:13:54 crc kubenswrapper[4715]: I1204 14:13:54.881557 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:13:55 crc kubenswrapper[4715]: I1204 14:13:55.074886 4715 generic.go:334] "Generic (PLEG): container finished" podID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerID="ba9b58b009e79708080733cec29056ee2ac6611d67a0d3fbdaa53df1aaf5d6fc" exitCode=0 Dec 04 14:13:55 crc kubenswrapper[4715]: I1204 14:13:55.074956 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerDied","Data":"ba9b58b009e79708080733cec29056ee2ac6611d67a0d3fbdaa53df1aaf5d6fc"} Dec 04 14:13:55 crc kubenswrapper[4715]: I1204 14:13:55.074988 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerStarted","Data":"2cd131d00ba7b0e6ed891d94173b11ed89fec8a284c7fb98452e658546683dcd"} Dec 04 14:13:55 crc kubenswrapper[4715]: I1204 14:13:55.078907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerStarted","Data":"45b3a87f2331d441b48ea1ba493ae15ca2c1f14000b3ac1f9cd590ef07d02b3b"} Dec 04 14:13:55 crc kubenswrapper[4715]: I1204 14:13:55.120709 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-52hfz" podStartSLOduration=3.711113507 podStartE2EDuration="6.120686515s" podCreationTimestamp="2025-12-04 14:13:49 +0000 UTC" firstStartedPulling="2025-12-04 14:13:51.043470959 +0000 UTC m=+988.112189174" lastFinishedPulling="2025-12-04 14:13:53.453043967 +0000 UTC m=+990.521762182" observedRunningTime="2025-12-04 14:13:55.117137351 +0000 UTC m=+992.185855576" watchObservedRunningTime="2025-12-04 14:13:55.120686515 +0000 UTC m=+992.189404730" Dec 04 14:13:55 crc kubenswrapper[4715]: I1204 14:13:55.480050 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj"] Dec 04 14:13:56 crc kubenswrapper[4715]: I1204 14:13:56.087708 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" event={"ID":"21b74fe1-833a-4744-b868-6b09c3e93458","Type":"ContainerStarted","Data":"157ad19cad994a35be4e659c940523a733d9f6499e91c86546a96cffb7938263"} Dec 04 14:13:57 crc kubenswrapper[4715]: I1204 14:13:57.098827 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerStarted","Data":"dd5ccf05890b09ab1746e9c667a5e13b5b4869ed198459f879bcbbb3808dfbd5"} Dec 04 14:13:58 crc kubenswrapper[4715]: I1204 14:13:58.413869 4715 generic.go:334] "Generic (PLEG): container finished" podID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerID="dd5ccf05890b09ab1746e9c667a5e13b5b4869ed198459f879bcbbb3808dfbd5" exitCode=0 Dec 04 14:13:58 crc kubenswrapper[4715]: I1204 14:13:58.413952 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerDied","Data":"dd5ccf05890b09ab1746e9c667a5e13b5b4869ed198459f879bcbbb3808dfbd5"} Dec 04 14:13:59 crc kubenswrapper[4715]: I1204 14:13:59.780159 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:59 crc kubenswrapper[4715]: I1204 14:13:59.781424 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:13:59 crc kubenswrapper[4715]: I1204 14:13:59.937235 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:14:00 crc kubenswrapper[4715]: I1204 14:14:00.811051 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:14:03 crc kubenswrapper[4715]: I1204 14:14:03.630788 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-52hfz"] Dec 04 14:14:03 crc kubenswrapper[4715]: I1204 14:14:03.631202 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-52hfz" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="registry-server" containerID="cri-o://45b3a87f2331d441b48ea1ba493ae15ca2c1f14000b3ac1f9cd590ef07d02b3b" gracePeriod=2 Dec 04 14:14:04 crc kubenswrapper[4715]: I1204 14:14:04.572946 4715 generic.go:334] "Generic (PLEG): container finished" podID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerID="45b3a87f2331d441b48ea1ba493ae15ca2c1f14000b3ac1f9cd590ef07d02b3b" exitCode=0 Dec 04 14:14:04 crc kubenswrapper[4715]: I1204 14:14:04.573312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerDied","Data":"45b3a87f2331d441b48ea1ba493ae15ca2c1f14000b3ac1f9cd590ef07d02b3b"} Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.466904 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.570308 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdrsc\" (UniqueName: \"kubernetes.io/projected/6b2f3a01-4232-4457-8034-c18cad1517cf-kube-api-access-hdrsc\") pod \"6b2f3a01-4232-4457-8034-c18cad1517cf\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.570529 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-catalog-content\") pod \"6b2f3a01-4232-4457-8034-c18cad1517cf\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.570568 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-utilities\") pod \"6b2f3a01-4232-4457-8034-c18cad1517cf\" (UID: \"6b2f3a01-4232-4457-8034-c18cad1517cf\") " Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.572326 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-utilities" (OuterVolumeSpecName: "utilities") pod "6b2f3a01-4232-4457-8034-c18cad1517cf" (UID: "6b2f3a01-4232-4457-8034-c18cad1517cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.578675 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2f3a01-4232-4457-8034-c18cad1517cf-kube-api-access-hdrsc" (OuterVolumeSpecName: "kube-api-access-hdrsc") pod "6b2f3a01-4232-4457-8034-c18cad1517cf" (UID: "6b2f3a01-4232-4457-8034-c18cad1517cf"). InnerVolumeSpecName "kube-api-access-hdrsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.641409 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52hfz" event={"ID":"6b2f3a01-4232-4457-8034-c18cad1517cf","Type":"ContainerDied","Data":"578490950a822360f0079b8e301234bb4b1d05612ac2159a24cb3b56cf8261a5"} Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.641473 4715 scope.go:117] "RemoveContainer" containerID="45b3a87f2331d441b48ea1ba493ae15ca2c1f14000b3ac1f9cd590ef07d02b3b" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.641654 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52hfz" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.651502 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b2f3a01-4232-4457-8034-c18cad1517cf" (UID: "6b2f3a01-4232-4457-8034-c18cad1517cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.672737 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.672788 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f3a01-4232-4457-8034-c18cad1517cf-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.672800 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdrsc\" (UniqueName: \"kubernetes.io/projected/6b2f3a01-4232-4457-8034-c18cad1517cf-kube-api-access-hdrsc\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.685917 4715 scope.go:117] "RemoveContainer" containerID="c5ede22eeefd8f833e476aab4304a13962c21cf4d2b9630c16014c680f3679de" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.744484 4715 scope.go:117] "RemoveContainer" containerID="4363dbe6fcd80ba00ed73da67c1fbfaf3201e0ec5ebd312a1218fc3dd4fad012" Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.977908 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-52hfz"] Dec 04 14:14:06 crc kubenswrapper[4715]: I1204 14:14:06.983420 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-52hfz"] Dec 04 14:14:07 crc kubenswrapper[4715]: I1204 14:14:07.191029 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" path="/var/lib/kubelet/pods/6b2f3a01-4232-4457-8034-c18cad1517cf/volumes" Dec 04 14:14:07 crc kubenswrapper[4715]: I1204 14:14:07.651387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" event={"ID":"21b74fe1-833a-4744-b868-6b09c3e93458","Type":"ContainerStarted","Data":"91ceab551b3baf79df03f076bdac3abe6dfdc9eefb57b1b1dbccca7aa120f6e6"} Dec 04 14:14:07 crc kubenswrapper[4715]: I1204 14:14:07.651563 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:14:07 crc kubenswrapper[4715]: I1204 14:14:07.654341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerStarted","Data":"a8fcd2f15b1d5953ffff8255924799c91258db4ec937ddbed78ee17cd95091e2"} Dec 04 14:14:07 crc kubenswrapper[4715]: I1204 14:14:07.686354 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" podStartSLOduration=2.6273924859999997 podStartE2EDuration="13.686333213s" podCreationTimestamp="2025-12-04 14:13:54 +0000 UTC" firstStartedPulling="2025-12-04 14:13:55.491502762 +0000 UTC m=+992.560220977" lastFinishedPulling="2025-12-04 14:14:06.550443499 +0000 UTC m=+1003.619161704" observedRunningTime="2025-12-04 14:14:07.682223963 +0000 UTC m=+1004.750942178" watchObservedRunningTime="2025-12-04 14:14:07.686333213 +0000 UTC m=+1004.755051428" Dec 04 14:14:07 crc kubenswrapper[4715]: I1204 14:14:07.703336 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hzf2f" podStartSLOduration=3.247865588 podStartE2EDuration="14.703311565s" podCreationTimestamp="2025-12-04 14:13:53 +0000 UTC" firstStartedPulling="2025-12-04 14:13:55.076663373 +0000 UTC m=+992.145381588" lastFinishedPulling="2025-12-04 14:14:06.53210936 +0000 UTC m=+1003.600827565" observedRunningTime="2025-12-04 14:14:07.7005152 +0000 UTC m=+1004.769233425" watchObservedRunningTime="2025-12-04 14:14:07.703311565 +0000 UTC m=+1004.772029790" Dec 04 14:14:08 crc kubenswrapper[4715]: I1204 14:14:08.759080 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:14:08 crc kubenswrapper[4715]: I1204 14:14:08.759154 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.439692 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-727fc"] Dec 04 14:14:11 crc kubenswrapper[4715]: E1204 14:14:11.440782 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="extract-utilities" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.440904 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="extract-utilities" Dec 04 14:14:11 crc kubenswrapper[4715]: E1204 14:14:11.440977 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="extract-content" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.441057 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="extract-content" Dec 04 14:14:11 crc kubenswrapper[4715]: E1204 14:14:11.441204 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="registry-server" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.441297 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="registry-server" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.441484 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2f3a01-4232-4457-8034-c18cad1517cf" containerName="registry-server" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.443265 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.454119 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-727fc"] Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.555130 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psx58\" (UniqueName: \"kubernetes.io/projected/dfb272da-8fb6-403f-9186-e90597ed07ee-kube-api-access-psx58\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.555249 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-utilities\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.555279 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-catalog-content\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.656527 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-utilities\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.656573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-catalog-content\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.656635 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psx58\" (UniqueName: \"kubernetes.io/projected/dfb272da-8fb6-403f-9186-e90597ed07ee-kube-api-access-psx58\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.657108 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-utilities\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.657141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-catalog-content\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.677115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psx58\" (UniqueName: \"kubernetes.io/projected/dfb272da-8fb6-403f-9186-e90597ed07ee-kube-api-access-psx58\") pod \"certified-operators-727fc\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:11 crc kubenswrapper[4715]: I1204 14:14:11.772854 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:12 crc kubenswrapper[4715]: I1204 14:14:12.043773 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-727fc"] Dec 04 14:14:12 crc kubenswrapper[4715]: W1204 14:14:12.050277 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfb272da_8fb6_403f_9186_e90597ed07ee.slice/crio-2de2f3c9bce17b01b6993792b3c1557c20ffdf11a9e0a339767ad89805f59db6 WatchSource:0}: Error finding container 2de2f3c9bce17b01b6993792b3c1557c20ffdf11a9e0a339767ad89805f59db6: Status 404 returned error can't find the container with id 2de2f3c9bce17b01b6993792b3c1557c20ffdf11a9e0a339767ad89805f59db6 Dec 04 14:14:12 crc kubenswrapper[4715]: I1204 14:14:12.684052 4715 generic.go:334] "Generic (PLEG): container finished" podID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerID="3ae78f12ddfe22a06cdf79596f395bf7fd173c9783771f286cad1f52d13a795f" exitCode=0 Dec 04 14:14:12 crc kubenswrapper[4715]: I1204 14:14:12.684200 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-727fc" event={"ID":"dfb272da-8fb6-403f-9186-e90597ed07ee","Type":"ContainerDied","Data":"3ae78f12ddfe22a06cdf79596f395bf7fd173c9783771f286cad1f52d13a795f"} Dec 04 14:14:12 crc kubenswrapper[4715]: I1204 14:14:12.684350 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-727fc" event={"ID":"dfb272da-8fb6-403f-9186-e90597ed07ee","Type":"ContainerStarted","Data":"2de2f3c9bce17b01b6993792b3c1557c20ffdf11a9e0a339767ad89805f59db6"} Dec 04 14:14:13 crc kubenswrapper[4715]: I1204 14:14:13.564499 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:14:13 crc kubenswrapper[4715]: I1204 14:14:13.564870 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:14:13 crc kubenswrapper[4715]: I1204 14:14:13.604173 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:14:13 crc kubenswrapper[4715]: I1204 14:14:13.725985 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:14:14 crc kubenswrapper[4715]: I1204 14:14:14.698436 4715 generic.go:334] "Generic (PLEG): container finished" podID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerID="240713f1e186e520910b881e1f5982b00e9ddbd2150806fe394a9ab25bfbb7c6" exitCode=0 Dec 04 14:14:14 crc kubenswrapper[4715]: I1204 14:14:14.698668 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-727fc" event={"ID":"dfb272da-8fb6-403f-9186-e90597ed07ee","Type":"ContainerDied","Data":"240713f1e186e520910b881e1f5982b00e9ddbd2150806fe394a9ab25bfbb7c6"} Dec 04 14:14:14 crc kubenswrapper[4715]: I1204 14:14:14.884965 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6cf7bc7f5d-kzpwj" Dec 04 14:14:15 crc kubenswrapper[4715]: I1204 14:14:15.811512 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-727fc" event={"ID":"dfb272da-8fb6-403f-9186-e90597ed07ee","Type":"ContainerStarted","Data":"0287c672c52b5c10c58575b6135adb3f519bf343a6073f554a90c5270fd2664b"} Dec 04 14:14:15 crc kubenswrapper[4715]: I1204 14:14:15.927101 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-727fc" podStartSLOduration=2.5067964910000002 podStartE2EDuration="4.927078665s" podCreationTimestamp="2025-12-04 14:14:11 +0000 UTC" firstStartedPulling="2025-12-04 14:14:12.685264329 +0000 UTC m=+1009.753982534" lastFinishedPulling="2025-12-04 14:14:15.105546493 +0000 UTC m=+1012.174264708" observedRunningTime="2025-12-04 14:14:15.923379076 +0000 UTC m=+1012.992097301" watchObservedRunningTime="2025-12-04 14:14:15.927078665 +0000 UTC m=+1012.995796900" Dec 04 14:14:16 crc kubenswrapper[4715]: I1204 14:14:16.028191 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzf2f"] Dec 04 14:14:16 crc kubenswrapper[4715]: I1204 14:14:16.028407 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hzf2f" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="registry-server" containerID="cri-o://a8fcd2f15b1d5953ffff8255924799c91258db4ec937ddbed78ee17cd95091e2" gracePeriod=2 Dec 04 14:14:16 crc kubenswrapper[4715]: I1204 14:14:16.825017 4715 generic.go:334] "Generic (PLEG): container finished" podID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerID="a8fcd2f15b1d5953ffff8255924799c91258db4ec937ddbed78ee17cd95091e2" exitCode=0 Dec 04 14:14:16 crc kubenswrapper[4715]: I1204 14:14:16.825073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerDied","Data":"a8fcd2f15b1d5953ffff8255924799c91258db4ec937ddbed78ee17cd95091e2"} Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.098339 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.157634 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hptzt\" (UniqueName: \"kubernetes.io/projected/cbf18951-ee68-4717-93c0-ec9a1826e37c-kube-api-access-hptzt\") pod \"cbf18951-ee68-4717-93c0-ec9a1826e37c\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.157818 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-catalog-content\") pod \"cbf18951-ee68-4717-93c0-ec9a1826e37c\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.157854 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-utilities\") pod \"cbf18951-ee68-4717-93c0-ec9a1826e37c\" (UID: \"cbf18951-ee68-4717-93c0-ec9a1826e37c\") " Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.158686 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-utilities" (OuterVolumeSpecName: "utilities") pod "cbf18951-ee68-4717-93c0-ec9a1826e37c" (UID: "cbf18951-ee68-4717-93c0-ec9a1826e37c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.163682 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf18951-ee68-4717-93c0-ec9a1826e37c-kube-api-access-hptzt" (OuterVolumeSpecName: "kube-api-access-hptzt") pod "cbf18951-ee68-4717-93c0-ec9a1826e37c" (UID: "cbf18951-ee68-4717-93c0-ec9a1826e37c"). InnerVolumeSpecName "kube-api-access-hptzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.174756 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbf18951-ee68-4717-93c0-ec9a1826e37c" (UID: "cbf18951-ee68-4717-93c0-ec9a1826e37c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.259521 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.259558 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbf18951-ee68-4717-93c0-ec9a1826e37c-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.259568 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hptzt\" (UniqueName: \"kubernetes.io/projected/cbf18951-ee68-4717-93c0-ec9a1826e37c-kube-api-access-hptzt\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.834967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzf2f" event={"ID":"cbf18951-ee68-4717-93c0-ec9a1826e37c","Type":"ContainerDied","Data":"2cd131d00ba7b0e6ed891d94173b11ed89fec8a284c7fb98452e658546683dcd"} Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.835086 4715 scope.go:117] "RemoveContainer" containerID="a8fcd2f15b1d5953ffff8255924799c91258db4ec937ddbed78ee17cd95091e2" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.835115 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzf2f" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.854595 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzf2f"] Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.858626 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzf2f"] Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.860987 4715 scope.go:117] "RemoveContainer" containerID="dd5ccf05890b09ab1746e9c667a5e13b5b4869ed198459f879bcbbb3808dfbd5" Dec 04 14:14:17 crc kubenswrapper[4715]: I1204 14:14:17.880848 4715 scope.go:117] "RemoveContainer" containerID="ba9b58b009e79708080733cec29056ee2ac6611d67a0d3fbdaa53df1aaf5d6fc" Dec 04 14:14:19 crc kubenswrapper[4715]: I1204 14:14:19.187771 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" path="/var/lib/kubelet/pods/cbf18951-ee68-4717-93c0-ec9a1826e37c/volumes" Dec 04 14:14:21 crc kubenswrapper[4715]: I1204 14:14:21.773972 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:21 crc kubenswrapper[4715]: I1204 14:14:21.775891 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:21 crc kubenswrapper[4715]: I1204 14:14:21.821597 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:21 crc kubenswrapper[4715]: I1204 14:14:21.901788 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:22 crc kubenswrapper[4715]: I1204 14:14:22.827972 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-727fc"] Dec 04 14:14:23 crc kubenswrapper[4715]: I1204 14:14:23.873628 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-727fc" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="registry-server" containerID="cri-o://0287c672c52b5c10c58575b6135adb3f519bf343a6073f554a90c5270fd2664b" gracePeriod=2 Dec 04 14:14:25 crc kubenswrapper[4715]: I1204 14:14:25.887927 4715 generic.go:334] "Generic (PLEG): container finished" podID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerID="0287c672c52b5c10c58575b6135adb3f519bf343a6073f554a90c5270fd2664b" exitCode=0 Dec 04 14:14:25 crc kubenswrapper[4715]: I1204 14:14:25.888070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-727fc" event={"ID":"dfb272da-8fb6-403f-9186-e90597ed07ee","Type":"ContainerDied","Data":"0287c672c52b5c10c58575b6135adb3f519bf343a6073f554a90c5270fd2664b"} Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.119121 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.187416 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-utilities\") pod \"dfb272da-8fb6-403f-9186-e90597ed07ee\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.187583 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-catalog-content\") pod \"dfb272da-8fb6-403f-9186-e90597ed07ee\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.187627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psx58\" (UniqueName: \"kubernetes.io/projected/dfb272da-8fb6-403f-9186-e90597ed07ee-kube-api-access-psx58\") pod \"dfb272da-8fb6-403f-9186-e90597ed07ee\" (UID: \"dfb272da-8fb6-403f-9186-e90597ed07ee\") " Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.189277 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-utilities" (OuterVolumeSpecName: "utilities") pod "dfb272da-8fb6-403f-9186-e90597ed07ee" (UID: "dfb272da-8fb6-403f-9186-e90597ed07ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.205375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb272da-8fb6-403f-9186-e90597ed07ee-kube-api-access-psx58" (OuterVolumeSpecName: "kube-api-access-psx58") pod "dfb272da-8fb6-403f-9186-e90597ed07ee" (UID: "dfb272da-8fb6-403f-9186-e90597ed07ee"). InnerVolumeSpecName "kube-api-access-psx58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.244135 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfb272da-8fb6-403f-9186-e90597ed07ee" (UID: "dfb272da-8fb6-403f-9186-e90597ed07ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.289136 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.289198 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psx58\" (UniqueName: \"kubernetes.io/projected/dfb272da-8fb6-403f-9186-e90597ed07ee-kube-api-access-psx58\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.289215 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb272da-8fb6-403f-9186-e90597ed07ee-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.899936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-727fc" event={"ID":"dfb272da-8fb6-403f-9186-e90597ed07ee","Type":"ContainerDied","Data":"2de2f3c9bce17b01b6993792b3c1557c20ffdf11a9e0a339767ad89805f59db6"} Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.900085 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-727fc" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.900486 4715 scope.go:117] "RemoveContainer" containerID="0287c672c52b5c10c58575b6135adb3f519bf343a6073f554a90c5270fd2664b" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.923422 4715 scope.go:117] "RemoveContainer" containerID="240713f1e186e520910b881e1f5982b00e9ddbd2150806fe394a9ab25bfbb7c6" Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.943315 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-727fc"] Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.948145 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-727fc"] Dec 04 14:14:26 crc kubenswrapper[4715]: I1204 14:14:26.962186 4715 scope.go:117] "RemoveContainer" containerID="3ae78f12ddfe22a06cdf79596f395bf7fd173c9783771f286cad1f52d13a795f" Dec 04 14:14:27 crc kubenswrapper[4715]: I1204 14:14:27.190390 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" path="/var/lib/kubelet/pods/dfb272da-8fb6-403f-9186-e90597ed07ee/volumes" Dec 04 14:14:38 crc kubenswrapper[4715]: I1204 14:14:38.758863 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:14:38 crc kubenswrapper[4715]: I1204 14:14:38.760284 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.201759 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf"] Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.203636 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="extract-content" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.203741 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="extract-content" Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.203832 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="extract-content" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.203907 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="extract-content" Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.203979 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="registry-server" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.204101 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="registry-server" Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.204189 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="extract-utilities" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.204266 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="extract-utilities" Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.204343 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="registry-server" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.204419 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="registry-server" Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.204497 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="extract-utilities" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.204603 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="extract-utilities" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.204840 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf18951-ee68-4717-93c0-ec9a1826e37c" containerName="registry-server" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.204947 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb272da-8fb6-403f-9186-e90597ed07ee" containerName="registry-server" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.205846 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.216755 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.217656 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-4gvmm" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.239163 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.240148 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.243181 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-zr2cn" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.248098 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.249241 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.251254 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6jx5k" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.254627 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.274369 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.280453 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.285135 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.289184 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-4bxqs" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.300738 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.302594 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.304656 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vxmcv" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.307839 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.315019 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.323144 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.324122 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.326467 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.327444 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.332311 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.332916 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pq9cn" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.333146 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-xn44b" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.339101 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.340442 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.345461 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-786r7" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.345674 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.350755 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.366268 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.367672 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.371570 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.373921 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rt5vb" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.402257 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xll8h\" (UniqueName: \"kubernetes.io/projected/fbc62f1d-2aa8-4b6b-8693-a16c349461d3-kube-api-access-xll8h\") pod \"heat-operator-controller-manager-5f64f6f8bb-gv5kg\" (UID: \"fbc62f1d-2aa8-4b6b-8693-a16c349461d3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.402305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qdcp\" (UniqueName: \"kubernetes.io/projected/4b1a2407-c585-4f22-95db-a2331a69623d-kube-api-access-4qdcp\") pod \"designate-operator-controller-manager-78b4bc895b-5m69r\" (UID: \"4b1a2407-c585-4f22-95db-a2331a69623d\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.402339 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vj9q\" (UniqueName: \"kubernetes.io/projected/6efc64e6-15fa-4c60-9a4e-18d326e4c759-kube-api-access-9vj9q\") pod \"cinder-operator-controller-manager-859b6ccc6-fdgb9\" (UID: \"6efc64e6-15fa-4c60-9a4e-18d326e4c759\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.402398 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgq9k\" (UniqueName: \"kubernetes.io/projected/96aa918a-abfb-44f6-bfa3-9849ef0f51e4-kube-api-access-cgq9k\") pod \"glance-operator-controller-manager-77987cd8cd-grcg8\" (UID: \"96aa918a-abfb-44f6-bfa3-9849ef0f51e4\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.402441 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhv72\" (UniqueName: \"kubernetes.io/projected/d7fc7c62-f1ba-4fcc-9b82-b980988c90cf-kube-api-access-lhv72\") pod \"barbican-operator-controller-manager-7d9dfd778-snfrf\" (UID: \"d7fc7c62-f1ba-4fcc-9b82-b980988c90cf\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.404637 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.443111 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.444406 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.447896 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-l6zmn" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.462098 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.463183 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.466611 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vgdkh" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.482218 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.483322 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.487541 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-4sw46" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528494 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4k58\" (UniqueName: \"kubernetes.io/projected/81510632-c76c-4465-8dec-d8d9c3e8c88d-kube-api-access-c4k58\") pod \"horizon-operator-controller-manager-68c6d99b8f-5glkm\" (UID: \"81510632-c76c-4465-8dec-d8d9c3e8c88d\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgq9k\" (UniqueName: \"kubernetes.io/projected/96aa918a-abfb-44f6-bfa3-9849ef0f51e4-kube-api-access-cgq9k\") pod \"glance-operator-controller-manager-77987cd8cd-grcg8\" (UID: \"96aa918a-abfb-44f6-bfa3-9849ef0f51e4\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528623 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhv72\" (UniqueName: \"kubernetes.io/projected/d7fc7c62-f1ba-4fcc-9b82-b980988c90cf-kube-api-access-lhv72\") pod \"barbican-operator-controller-manager-7d9dfd778-snfrf\" (UID: \"d7fc7c62-f1ba-4fcc-9b82-b980988c90cf\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djpbp\" (UniqueName: \"kubernetes.io/projected/ec0c61ca-d902-4703-85bc-87b55736fe78-kube-api-access-djpbp\") pod \"ironic-operator-controller-manager-6c548fd776-cw997\" (UID: \"ec0c61ca-d902-4703-85bc-87b55736fe78\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6xnt\" (UniqueName: \"kubernetes.io/projected/9c67b2dc-8ac4-4dd0-af9f-19e55d368092-kube-api-access-v6xnt\") pod \"keystone-operator-controller-manager-7765d96ddf-rtj2r\" (UID: \"9c67b2dc-8ac4-4dd0-af9f-19e55d368092\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528755 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6l6c\" (UniqueName: \"kubernetes.io/projected/8a9a0373-b39e-496a-aba7-456bddcf3a65-kube-api-access-v6l6c\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xll8h\" (UniqueName: \"kubernetes.io/projected/fbc62f1d-2aa8-4b6b-8693-a16c349461d3-kube-api-access-xll8h\") pod \"heat-operator-controller-manager-5f64f6f8bb-gv5kg\" (UID: \"fbc62f1d-2aa8-4b6b-8693-a16c349461d3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qdcp\" (UniqueName: \"kubernetes.io/projected/4b1a2407-c585-4f22-95db-a2331a69623d-kube-api-access-4qdcp\") pod \"designate-operator-controller-manager-78b4bc895b-5m69r\" (UID: \"4b1a2407-c585-4f22-95db-a2331a69623d\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.528860 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vj9q\" (UniqueName: \"kubernetes.io/projected/6efc64e6-15fa-4c60-9a4e-18d326e4c759-kube-api-access-9vj9q\") pod \"cinder-operator-controller-manager-859b6ccc6-fdgb9\" (UID: \"6efc64e6-15fa-4c60-9a4e-18d326e4c759\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.541891 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.549656 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.573112 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.576340 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.576763 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.577757 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.584813 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ql69z" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.585324 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xsd9k" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.586003 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vj9q\" (UniqueName: \"kubernetes.io/projected/6efc64e6-15fa-4c60-9a4e-18d326e4c759-kube-api-access-9vj9q\") pod \"cinder-operator-controller-manager-859b6ccc6-fdgb9\" (UID: \"6efc64e6-15fa-4c60-9a4e-18d326e4c759\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.600713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qdcp\" (UniqueName: \"kubernetes.io/projected/4b1a2407-c585-4f22-95db-a2331a69623d-kube-api-access-4qdcp\") pod \"designate-operator-controller-manager-78b4bc895b-5m69r\" (UID: \"4b1a2407-c585-4f22-95db-a2331a69623d\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.603674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xll8h\" (UniqueName: \"kubernetes.io/projected/fbc62f1d-2aa8-4b6b-8693-a16c349461d3-kube-api-access-xll8h\") pod \"heat-operator-controller-manager-5f64f6f8bb-gv5kg\" (UID: \"fbc62f1d-2aa8-4b6b-8693-a16c349461d3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.608777 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgq9k\" (UniqueName: \"kubernetes.io/projected/96aa918a-abfb-44f6-bfa3-9849ef0f51e4-kube-api-access-cgq9k\") pod \"glance-operator-controller-manager-77987cd8cd-grcg8\" (UID: \"96aa918a-abfb-44f6-bfa3-9849ef0f51e4\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.611297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.613941 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhv72\" (UniqueName: \"kubernetes.io/projected/d7fc7c62-f1ba-4fcc-9b82-b980988c90cf-kube-api-access-lhv72\") pod \"barbican-operator-controller-manager-7d9dfd778-snfrf\" (UID: \"d7fc7c62-f1ba-4fcc-9b82-b980988c90cf\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629658 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbjf4\" (UniqueName: \"kubernetes.io/projected/5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd-kube-api-access-hbjf4\") pod \"mariadb-operator-controller-manager-56bbcc9d85-xncpr\" (UID: \"5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4k58\" (UniqueName: \"kubernetes.io/projected/81510632-c76c-4465-8dec-d8d9c3e8c88d-kube-api-access-c4k58\") pod \"horizon-operator-controller-manager-68c6d99b8f-5glkm\" (UID: \"81510632-c76c-4465-8dec-d8d9c3e8c88d\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629771 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnsxm\" (UniqueName: \"kubernetes.io/projected/45a4ad9a-e82a-466a-954b-12127626b577-kube-api-access-nnsxm\") pod \"manila-operator-controller-manager-7c79b5df47-hhjwx\" (UID: \"45a4ad9a-e82a-466a-954b-12127626b577\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629802 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwxpb\" (UniqueName: \"kubernetes.io/projected/3521ae52-a647-4190-8bbd-1100bb31dae5-kube-api-access-dwxpb\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-6hbp2\" (UID: \"3521ae52-a647-4190-8bbd-1100bb31dae5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629829 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629854 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djpbp\" (UniqueName: \"kubernetes.io/projected/ec0c61ca-d902-4703-85bc-87b55736fe78-kube-api-access-djpbp\") pod \"ironic-operator-controller-manager-6c548fd776-cw997\" (UID: \"ec0c61ca-d902-4703-85bc-87b55736fe78\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629873 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6xnt\" (UniqueName: \"kubernetes.io/projected/9c67b2dc-8ac4-4dd0-af9f-19e55d368092-kube-api-access-v6xnt\") pod \"keystone-operator-controller-manager-7765d96ddf-rtj2r\" (UID: \"9c67b2dc-8ac4-4dd0-af9f-19e55d368092\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.629893 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6l6c\" (UniqueName: \"kubernetes.io/projected/8a9a0373-b39e-496a-aba7-456bddcf3a65-kube-api-access-v6l6c\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.630239 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:44 crc kubenswrapper[4715]: E1204 14:14:44.630279 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert podName:8a9a0373-b39e-496a-aba7-456bddcf3a65 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:45.130262863 +0000 UTC m=+1042.198981078 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert") pod "infra-operator-controller-manager-57548d458d-f8cgr" (UID: "8a9a0373-b39e-496a-aba7-456bddcf3a65") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.630676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.644894 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.679300 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.690481 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.697098 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.701566 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.710537 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-k28zf" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.728453 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.731347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnsxm\" (UniqueName: \"kubernetes.io/projected/45a4ad9a-e82a-466a-954b-12127626b577-kube-api-access-nnsxm\") pod \"manila-operator-controller-manager-7c79b5df47-hhjwx\" (UID: \"45a4ad9a-e82a-466a-954b-12127626b577\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.731413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwxpb\" (UniqueName: \"kubernetes.io/projected/3521ae52-a647-4190-8bbd-1100bb31dae5-kube-api-access-dwxpb\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-6hbp2\" (UID: \"3521ae52-a647-4190-8bbd-1100bb31dae5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.731446 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p98ww\" (UniqueName: \"kubernetes.io/projected/ab0e965e-1d56-4b31-8cf1-6a21892d38ca-kube-api-access-p98ww\") pod \"nova-operator-controller-manager-697bc559fc-vnndg\" (UID: \"ab0e965e-1d56-4b31-8cf1-6a21892d38ca\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.731535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kdqr\" (UniqueName: \"kubernetes.io/projected/4238708b-a19f-485f-a8a7-66b58aac168f-kube-api-access-6kdqr\") pod \"octavia-operator-controller-manager-998648c74-dg7m6\" (UID: \"4238708b-a19f-485f-a8a7-66b58aac168f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.731562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbjf4\" (UniqueName: \"kubernetes.io/projected/5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd-kube-api-access-hbjf4\") pod \"mariadb-operator-controller-manager-56bbcc9d85-xncpr\" (UID: \"5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.732487 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4k58\" (UniqueName: \"kubernetes.io/projected/81510632-c76c-4465-8dec-d8d9c3e8c88d-kube-api-access-c4k58\") pod \"horizon-operator-controller-manager-68c6d99b8f-5glkm\" (UID: \"81510632-c76c-4465-8dec-d8d9c3e8c88d\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.738235 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6l6c\" (UniqueName: \"kubernetes.io/projected/8a9a0373-b39e-496a-aba7-456bddcf3a65-kube-api-access-v6l6c\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.752412 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.752475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djpbp\" (UniqueName: \"kubernetes.io/projected/ec0c61ca-d902-4703-85bc-87b55736fe78-kube-api-access-djpbp\") pod \"ironic-operator-controller-manager-6c548fd776-cw997\" (UID: \"ec0c61ca-d902-4703-85bc-87b55736fe78\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.753460 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.760283 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6xnt\" (UniqueName: \"kubernetes.io/projected/9c67b2dc-8ac4-4dd0-af9f-19e55d368092-kube-api-access-v6xnt\") pod \"keystone-operator-controller-manager-7765d96ddf-rtj2r\" (UID: \"9c67b2dc-8ac4-4dd0-af9f-19e55d368092\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.765400 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.765594 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ppzrl" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.773507 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnsxm\" (UniqueName: \"kubernetes.io/projected/45a4ad9a-e82a-466a-954b-12127626b577-kube-api-access-nnsxm\") pod \"manila-operator-controller-manager-7c79b5df47-hhjwx\" (UID: \"45a4ad9a-e82a-466a-954b-12127626b577\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.774685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbjf4\" (UniqueName: \"kubernetes.io/projected/5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd-kube-api-access-hbjf4\") pod \"mariadb-operator-controller-manager-56bbcc9d85-xncpr\" (UID: \"5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.787580 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwxpb\" (UniqueName: \"kubernetes.io/projected/3521ae52-a647-4190-8bbd-1100bb31dae5-kube-api-access-dwxpb\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-6hbp2\" (UID: \"3521ae52-a647-4190-8bbd-1100bb31dae5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.790217 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.838654 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.857753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kdqr\" (UniqueName: \"kubernetes.io/projected/4238708b-a19f-485f-a8a7-66b58aac168f-kube-api-access-6kdqr\") pod \"octavia-operator-controller-manager-998648c74-dg7m6\" (UID: \"4238708b-a19f-485f-a8a7-66b58aac168f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.857852 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk55c\" (UniqueName: \"kubernetes.io/projected/84876a00-902b-41f4-af69-f39d588e34fe-kube-api-access-lk55c\") pod \"ovn-operator-controller-manager-b6456fdb6-wwc57\" (UID: \"84876a00-902b-41f4-af69-f39d588e34fe\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.858005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p98ww\" (UniqueName: \"kubernetes.io/projected/ab0e965e-1d56-4b31-8cf1-6a21892d38ca-kube-api-access-p98ww\") pod \"nova-operator-controller-manager-697bc559fc-vnndg\" (UID: \"ab0e965e-1d56-4b31-8cf1-6a21892d38ca\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.860702 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.863856 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.866722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.870108 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-w2drg"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.871429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.886864 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.898402 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.905166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jh5tz" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.916188 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.917943 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.918824 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.921551 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-84h6d" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.921816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p98ww\" (UniqueName: \"kubernetes.io/projected/ab0e965e-1d56-4b31-8cf1-6a21892d38ca-kube-api-access-p98ww\") pod \"nova-operator-controller-manager-697bc559fc-vnndg\" (UID: \"ab0e965e-1d56-4b31-8cf1-6a21892d38ca\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.926099 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kdqr\" (UniqueName: \"kubernetes.io/projected/4238708b-a19f-485f-a8a7-66b58aac168f-kube-api-access-6kdqr\") pod \"octavia-operator-controller-manager-998648c74-dg7m6\" (UID: \"4238708b-a19f-485f-a8a7-66b58aac168f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.937840 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-w2drg"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.937915 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.954096 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5"] Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.961458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt6v6\" (UniqueName: \"kubernetes.io/projected/73189a74-45fd-4ea8-92ce-bfa038531159-kube-api-access-wt6v6\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.961540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk55c\" (UniqueName: \"kubernetes.io/projected/84876a00-902b-41f4-af69-f39d588e34fe-kube-api-access-lk55c\") pod \"ovn-operator-controller-manager-b6456fdb6-wwc57\" (UID: \"84876a00-902b-41f4-af69-f39d588e34fe\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.961572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:44 crc kubenswrapper[4715]: I1204 14:14:44.962090 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:44.981865 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-sg77c"] Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:44.985663 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jpfgr" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:44.988642 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:44.999974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.000527 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.008352 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-sg77c"] Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.039305 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-phmdb" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.052660 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn"] Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.053654 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn"] Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.053734 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.062257 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6w9z\" (UniqueName: \"kubernetes.io/projected/cec0926c-ac5c-4ef2-aefc-a81c7885d810-kube-api-access-q6w9z\") pod \"telemetry-operator-controller-manager-76cc84c6bb-xc2z5\" (UID: \"cec0926c-ac5c-4ef2-aefc-a81c7885d810\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.062357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt6v6\" (UniqueName: \"kubernetes.io/projected/73189a74-45fd-4ea8-92ce-bfa038531159-kube-api-access-wt6v6\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.062381 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl89v\" (UniqueName: \"kubernetes.io/projected/9d367397-9c5c-4ef7-bf95-91d0a453cf04-kube-api-access-jl89v\") pod \"placement-operator-controller-manager-78f8948974-w2drg\" (UID: \"9d367397-9c5c-4ef7-bf95-91d0a453cf04\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.062419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.062439 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4dg6\" (UniqueName: \"kubernetes.io/projected/6d5d27a6-eba8-4699-a415-e0e8c283d9cf-kube-api-access-z4dg6\") pod \"swift-operator-controller-manager-5f8c65bbfc-qq7m7\" (UID: \"6d5d27a6-eba8-4699-a415-e0e8c283d9cf\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.063248 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.063295 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert podName:73189a74-45fd-4ea8-92ce-bfa038531159 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:45.563277796 +0000 UTC m=+1042.631996011 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" (UID: "73189a74-45fd-4ea8-92ce-bfa038531159") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.077950 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.084823 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk55c\" (UniqueName: \"kubernetes.io/projected/84876a00-902b-41f4-af69-f39d588e34fe-kube-api-access-lk55c\") pod \"ovn-operator-controller-manager-b6456fdb6-wwc57\" (UID: \"84876a00-902b-41f4-af69-f39d588e34fe\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.085373 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.088247 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-vqwcp" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.200886 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd25j\" (UniqueName: \"kubernetes.io/projected/88013762-6dd2-4994-b34e-21c89ed5afee-kube-api-access-dd25j\") pod \"test-operator-controller-manager-5854674fcc-sg77c\" (UID: \"88013762-6dd2-4994-b34e-21c89ed5afee\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.200963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl89v\" (UniqueName: \"kubernetes.io/projected/9d367397-9c5c-4ef7-bf95-91d0a453cf04-kube-api-access-jl89v\") pod \"placement-operator-controller-manager-78f8948974-w2drg\" (UID: \"9d367397-9c5c-4ef7-bf95-91d0a453cf04\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.201031 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4dg6\" (UniqueName: \"kubernetes.io/projected/6d5d27a6-eba8-4699-a415-e0e8c283d9cf-kube-api-access-z4dg6\") pod \"swift-operator-controller-manager-5f8c65bbfc-qq7m7\" (UID: \"6d5d27a6-eba8-4699-a415-e0e8c283d9cf\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.212688 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9wsf\" (UniqueName: \"kubernetes.io/projected/d58523b1-200c-4b54-ab7c-fab0ee9ff571-kube-api-access-r9wsf\") pod \"watcher-operator-controller-manager-769dc69bc-8ztmn\" (UID: \"d58523b1-200c-4b54-ab7c-fab0ee9ff571\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.212788 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6w9z\" (UniqueName: \"kubernetes.io/projected/cec0926c-ac5c-4ef2-aefc-a81c7885d810-kube-api-access-q6w9z\") pod \"telemetry-operator-controller-manager-76cc84c6bb-xc2z5\" (UID: \"cec0926c-ac5c-4ef2-aefc-a81c7885d810\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.212910 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.213183 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.213248 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert podName:8a9a0373-b39e-496a-aba7-456bddcf3a65 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:46.21322908 +0000 UTC m=+1043.281947295 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert") pod "infra-operator-controller-manager-57548d458d-f8cgr" (UID: "8a9a0373-b39e-496a-aba7-456bddcf3a65") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.221210 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt6v6\" (UniqueName: \"kubernetes.io/projected/73189a74-45fd-4ea8-92ce-bfa038531159-kube-api-access-wt6v6\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.224600 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.226132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.313906 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9wsf\" (UniqueName: \"kubernetes.io/projected/d58523b1-200c-4b54-ab7c-fab0ee9ff571-kube-api-access-r9wsf\") pod \"watcher-operator-controller-manager-769dc69bc-8ztmn\" (UID: \"d58523b1-200c-4b54-ab7c-fab0ee9ff571\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.314151 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd25j\" (UniqueName: \"kubernetes.io/projected/88013762-6dd2-4994-b34e-21c89ed5afee-kube-api-access-dd25j\") pod \"test-operator-controller-manager-5854674fcc-sg77c\" (UID: \"88013762-6dd2-4994-b34e-21c89ed5afee\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.356023 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4dg6\" (UniqueName: \"kubernetes.io/projected/6d5d27a6-eba8-4699-a415-e0e8c283d9cf-kube-api-access-z4dg6\") pod \"swift-operator-controller-manager-5f8c65bbfc-qq7m7\" (UID: \"6d5d27a6-eba8-4699-a415-e0e8c283d9cf\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.360322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9wsf\" (UniqueName: \"kubernetes.io/projected/d58523b1-200c-4b54-ab7c-fab0ee9ff571-kube-api-access-r9wsf\") pod \"watcher-operator-controller-manager-769dc69bc-8ztmn\" (UID: \"d58523b1-200c-4b54-ab7c-fab0ee9ff571\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.362407 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd25j\" (UniqueName: \"kubernetes.io/projected/88013762-6dd2-4994-b34e-21c89ed5afee-kube-api-access-dd25j\") pod \"test-operator-controller-manager-5854674fcc-sg77c\" (UID: \"88013762-6dd2-4994-b34e-21c89ed5afee\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.363994 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl89v\" (UniqueName: \"kubernetes.io/projected/9d367397-9c5c-4ef7-bf95-91d0a453cf04-kube-api-access-jl89v\") pod \"placement-operator-controller-manager-78f8948974-w2drg\" (UID: \"9d367397-9c5c-4ef7-bf95-91d0a453cf04\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.373106 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6w9z\" (UniqueName: \"kubernetes.io/projected/cec0926c-ac5c-4ef2-aefc-a81c7885d810-kube-api-access-q6w9z\") pod \"telemetry-operator-controller-manager-76cc84c6bb-xc2z5\" (UID: \"cec0926c-ac5c-4ef2-aefc-a81c7885d810\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.376182 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.444366 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.540478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.568060 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.581337 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5"] Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.582487 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.601572 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-krhrd" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.601765 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.601877 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.651669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.652368 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.652560 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert podName:73189a74-45fd-4ea8-92ce-bfa038531159 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:46.652445659 +0000 UTC m=+1043.721163874 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" (UID: "73189a74-45fd-4ea8-92ce-bfa038531159") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.671643 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5"] Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.672866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.813371 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.813650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbsp\" (UniqueName: \"kubernetes.io/projected/00ef5230-a1c9-4609-ba71-e90760f29a15-kube-api-access-jsbsp\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.813773 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.914757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.914805 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbsp\" (UniqueName: \"kubernetes.io/projected/00ef5230-a1c9-4609-ba71-e90760f29a15-kube-api-access-jsbsp\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: I1204 14:14:45.914839 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.915020 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.915084 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:46.415068324 +0000 UTC m=+1043.483786539 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "webhook-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.915891 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:14:45 crc kubenswrapper[4715]: E1204 14:14:45.915975 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:46.415953907 +0000 UTC m=+1043.484672122 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "metrics-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.026711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbsp\" (UniqueName: \"kubernetes.io/projected/00ef5230-a1c9-4609-ba71-e90760f29a15-kube-api-access-jsbsp\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.052741 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592"] Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.053587 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592"] Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.053674 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.141008 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-pvzb5" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.212305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7csxr\" (UniqueName: \"kubernetes.io/projected/43170033-ddda-4fdf-b0e7-8d3b6d15e4f9-kube-api-access-7csxr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9g592\" (UID: \"43170033-ddda-4fdf-b0e7-8d3b6d15e4f9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.288232 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8"] Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.316053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.316206 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7csxr\" (UniqueName: \"kubernetes.io/projected/43170033-ddda-4fdf-b0e7-8d3b6d15e4f9-kube-api-access-7csxr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9g592\" (UID: \"43170033-ddda-4fdf-b0e7-8d3b6d15e4f9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.317609 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.317708 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert podName:8a9a0373-b39e-496a-aba7-456bddcf3a65 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:48.317685748 +0000 UTC m=+1045.386403963 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert") pod "infra-operator-controller-manager-57548d458d-f8cgr" (UID: "8a9a0373-b39e-496a-aba7-456bddcf3a65") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.399773 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg"] Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.419449 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.419880 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:47.419835678 +0000 UTC m=+1044.488553883 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "metrics-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.420054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7csxr\" (UniqueName: \"kubernetes.io/projected/43170033-ddda-4fdf-b0e7-8d3b6d15e4f9-kube-api-access-7csxr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9g592\" (UID: \"43170033-ddda-4fdf-b0e7-8d3b6d15e4f9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.422400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.422583 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.422954 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.423017 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:47.422999373 +0000 UTC m=+1044.491717578 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "webhook-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.437853 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" Dec 04 14:14:46 crc kubenswrapper[4715]: I1204 14:14:46.750796 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.751091 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:46 crc kubenswrapper[4715]: E1204 14:14:46.751194 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert podName:73189a74-45fd-4ea8-92ce-bfa038531159 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:48.751171984 +0000 UTC m=+1045.819890199 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" (UID: "73189a74-45fd-4ea8-92ce-bfa038531159") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:47 crc kubenswrapper[4715]: I1204 14:14:47.177821 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" event={"ID":"96aa918a-abfb-44f6-bfa3-9849ef0f51e4","Type":"ContainerStarted","Data":"2040190473a5af4656d9bfb09ed3e1fae33152cfd4860ca562d8ceb7079284c3"} Dec 04 14:14:47 crc kubenswrapper[4715]: I1204 14:14:47.180227 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" event={"ID":"fbc62f1d-2aa8-4b6b-8693-a16c349461d3","Type":"ContainerStarted","Data":"c233ab8514848a9c4f4d52cba60bd6327f044fdaa620a81ca07ca05738f5008d"} Dec 04 14:14:47 crc kubenswrapper[4715]: I1204 14:14:47.356678 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr"] Dec 04 14:14:47 crc kubenswrapper[4715]: I1204 14:14:47.453322 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:47 crc kubenswrapper[4715]: I1204 14:14:47.453376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:47 crc kubenswrapper[4715]: E1204 14:14:47.453473 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:14:47 crc kubenswrapper[4715]: E1204 14:14:47.453475 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:14:47 crc kubenswrapper[4715]: E1204 14:14:47.453527 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:49.453511369 +0000 UTC m=+1046.522229584 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "webhook-server-cert" not found Dec 04 14:14:47 crc kubenswrapper[4715]: E1204 14:14:47.453541 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:49.45353521 +0000 UTC m=+1046.522253425 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "metrics-server-cert" not found Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.212341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" event={"ID":"5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd","Type":"ContainerStarted","Data":"40693a9422bad3b4f5efd24e4022f4a0624f12de9696721172b04d6c25ddc16c"} Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.370723 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.370976 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.371044 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert podName:8a9a0373-b39e-496a-aba7-456bddcf3a65 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:52.371011807 +0000 UTC m=+1049.439730022 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert") pod "infra-operator-controller-manager-57548d458d-f8cgr" (UID: "8a9a0373-b39e-496a-aba7-456bddcf3a65") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.416593 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.447699 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.455834 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg"] Dec 04 14:14:48 crc kubenswrapper[4715]: W1204 14:14:48.473026 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d5d27a6_eba8_4699_a415_e0e8c283d9cf.slice/crio-3720004cb2e154b2e634420f83616b80916f20940534b383e4713870f49fc6a4 WatchSource:0}: Error finding container 3720004cb2e154b2e634420f83616b80916f20940534b383e4713870f49fc6a4: Status 404 returned error can't find the container with id 3720004cb2e154b2e634420f83616b80916f20940534b383e4713870f49fc6a4 Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.476724 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.502326 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.548158 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.594410 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.627214 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.633354 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.639949 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-w2drg"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.653874 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn"] Dec 04 14:14:48 crc kubenswrapper[4715]: W1204 14:14:48.658153 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4238708b_a19f_485f_a8a7_66b58aac168f.slice/crio-160b639fee827b1c2da27e9d51405bcd1df536c4c1e352f2d32652083377ae78 WatchSource:0}: Error finding container 160b639fee827b1c2da27e9d51405bcd1df536c4c1e352f2d32652083377ae78: Status 404 returned error can't find the container with id 160b639fee827b1c2da27e9d51405bcd1df536c4c1e352f2d32652083377ae78 Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.673118 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6"] Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.696744 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf"] Dec 04 14:14:48 crc kubenswrapper[4715]: W1204 14:14:48.698139 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7fc7c62_f1ba_4fcc_9b82_b980988c90cf.slice/crio-6f0a217e996bd18edb546cb0822096a744d619ee301fc79a6c3ec92873e8cda6 WatchSource:0}: Error finding container 6f0a217e996bd18edb546cb0822096a744d619ee301fc79a6c3ec92873e8cda6: Status 404 returned error can't find the container with id 6f0a217e996bd18edb546cb0822096a744d619ee301fc79a6c3ec92873e8cda6 Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.710640 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6kdqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-dg7m6_openstack-operators(4238708b-a19f-485f-a8a7-66b58aac168f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.715851 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592"] Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.717377 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6kdqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-dg7m6_openstack-operators(4238708b-a19f-485f-a8a7-66b58aac168f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.721394 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" podUID="4238708b-a19f-485f-a8a7-66b58aac168f" Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.729945 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997"] Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.735823 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9wsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-8ztmn_openstack-operators(d58523b1-200c-4b54-ab7c-fab0ee9ff571): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.735957 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lhv72,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-snfrf_openstack-operators(d7fc7c62-f1ba-4fcc-9b82-b980988c90cf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.750261 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7csxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-9g592_openstack-operators(43170033-ddda-4fdf-b0e7-8d3b6d15e4f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.750383 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lhv72,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-snfrf_openstack-operators(d7fc7c62-f1ba-4fcc-9b82-b980988c90cf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.751343 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" podUID="43170033-ddda-4fdf-b0e7-8d3b6d15e4f9" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.751624 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" podUID="d7fc7c62-f1ba-4fcc-9b82-b980988c90cf" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.753617 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-djpbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-cw997_openstack-operators(ec0c61ca-d902-4703-85bc-87b55736fe78): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.761631 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-djpbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-cw997_openstack-operators(ec0c61ca-d902-4703-85bc-87b55736fe78): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.762779 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" podUID="ec0c61ca-d902-4703-85bc-87b55736fe78" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.765647 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dd25j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-sg77c_openstack-operators(88013762-6dd2-4994-b34e-21c89ed5afee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.772334 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r"] Dec 04 14:14:48 crc kubenswrapper[4715]: W1204 14:14:48.777090 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b1a2407_c585_4f22_95db_a2331a69623d.slice/crio-a4be7d7a0d21be6de8471ef098f489d62a5e738296916df42b7898bdd1896706 WatchSource:0}: Error finding container a4be7d7a0d21be6de8471ef098f489d62a5e738296916df42b7898bdd1896706: Status 404 returned error can't find the container with id a4be7d7a0d21be6de8471ef098f489d62a5e738296916df42b7898bdd1896706 Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.779564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.779738 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.780531 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert podName:73189a74-45fd-4ea8-92ce-bfa038531159 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:52.780508624 +0000 UTC m=+1049.849226849 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" (UID: "73189a74-45fd-4ea8-92ce-bfa038531159") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.780399 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4qdcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-5m69r_openstack-operators(4b1a2407-c585-4f22-95db-a2331a69623d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.783490 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4qdcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-5m69r_openstack-operators(4b1a2407-c585-4f22-95db-a2331a69623d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.784906 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" podUID="4b1a2407-c585-4f22-95db-a2331a69623d" Dec 04 14:14:48 crc kubenswrapper[4715]: I1204 14:14:48.797616 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-sg77c"] Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.815620 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dd25j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-sg77c_openstack-operators(88013762-6dd2-4994-b34e-21c89ed5afee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:14:48 crc kubenswrapper[4715]: E1204 14:14:48.816912 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" podUID="88013762-6dd2-4994-b34e-21c89ed5afee" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.221704 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" event={"ID":"9d367397-9c5c-4ef7-bf95-91d0a453cf04","Type":"ContainerStarted","Data":"7a1f992b05cb5ea7584a6b6d53af58387f36dd81b3db098ed54abe4f9eb85b7b"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.224290 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" event={"ID":"9c67b2dc-8ac4-4dd0-af9f-19e55d368092","Type":"ContainerStarted","Data":"09e75e67fa06c89b025bde5abf79474426365a6c418edfd418d9d863c6663e00"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.226614 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" event={"ID":"cec0926c-ac5c-4ef2-aefc-a81c7885d810","Type":"ContainerStarted","Data":"2f8089f79c975928dc18efa767f7f874aca068e1f0def14af10dfe5398366a0f"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.228213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" event={"ID":"d58523b1-200c-4b54-ab7c-fab0ee9ff571","Type":"ContainerStarted","Data":"e8ec395698f0c838714d1e1a0b6150c1e94316574b21a9825fe9a42a551289d8"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.229429 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" event={"ID":"43170033-ddda-4fdf-b0e7-8d3b6d15e4f9","Type":"ContainerStarted","Data":"deae97f3d614b491250be1cb6554420c04a23843badc967cf6fd1aac91f701ef"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.241965 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" event={"ID":"45a4ad9a-e82a-466a-954b-12127626b577","Type":"ContainerStarted","Data":"98360835e9e3e68f44de93e9619b7ee516689984d9520d6c4d71c922870c3191"} Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.250078 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" podUID="43170033-ddda-4fdf-b0e7-8d3b6d15e4f9" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.250847 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" event={"ID":"d7fc7c62-f1ba-4fcc-9b82-b980988c90cf","Type":"ContainerStarted","Data":"6f0a217e996bd18edb546cb0822096a744d619ee301fc79a6c3ec92873e8cda6"} Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.258400 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" podUID="d7fc7c62-f1ba-4fcc-9b82-b980988c90cf" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.262811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" event={"ID":"81510632-c76c-4465-8dec-d8d9c3e8c88d","Type":"ContainerStarted","Data":"6afc648d48df0a6198e569b5d353e97a92e4329cc7d9db55627a804bdbd98da4"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.275219 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" event={"ID":"6d5d27a6-eba8-4699-a415-e0e8c283d9cf","Type":"ContainerStarted","Data":"3720004cb2e154b2e634420f83616b80916f20940534b383e4713870f49fc6a4"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.280381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" event={"ID":"88013762-6dd2-4994-b34e-21c89ed5afee","Type":"ContainerStarted","Data":"a6f350a3882a21f087edb68bb17318969a9c25645139919cc63da7d54e1e16a8"} Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.284805 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" podUID="88013762-6dd2-4994-b34e-21c89ed5afee" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.295848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" event={"ID":"6efc64e6-15fa-4c60-9a4e-18d326e4c759","Type":"ContainerStarted","Data":"11b1b5a64d5dfc002af61d4422e74b53a39f13f053284b041a14179b45b96a59"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.300691 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" event={"ID":"4238708b-a19f-485f-a8a7-66b58aac168f","Type":"ContainerStarted","Data":"160b639fee827b1c2da27e9d51405bcd1df536c4c1e352f2d32652083377ae78"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.303968 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" event={"ID":"ec0c61ca-d902-4703-85bc-87b55736fe78","Type":"ContainerStarted","Data":"3f06d10ce1bc4efa7bf6af1c8cec02d1f0f0f767fa3a5fc7d2b889e7d2451163"} Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.311538 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" podUID="4238708b-a19f-485f-a8a7-66b58aac168f" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.312381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" event={"ID":"84876a00-902b-41f4-af69-f39d588e34fe","Type":"ContainerStarted","Data":"5574e84375c4a9ca33b7ef72f29243fa9e904236e9b99468d83ea94197ddc283"} Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.313629 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" podUID="ec0c61ca-d902-4703-85bc-87b55736fe78" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.317949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" event={"ID":"3521ae52-a647-4190-8bbd-1100bb31dae5","Type":"ContainerStarted","Data":"7255d717ee43c8ae81288845128b5df5c115cfa48d42539677a80553dbc213e5"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.345605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" event={"ID":"ab0e965e-1d56-4b31-8cf1-6a21892d38ca","Type":"ContainerStarted","Data":"7d972e88310f1bc6e62fbe787cce3512b39047baa9f274b2c271809fa5dd2f2b"} Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.348847 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" event={"ID":"4b1a2407-c585-4f22-95db-a2331a69623d","Type":"ContainerStarted","Data":"a4be7d7a0d21be6de8471ef098f489d62a5e738296916df42b7898bdd1896706"} Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.353700 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" podUID="4b1a2407-c585-4f22-95db-a2331a69623d" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.495104 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:49 crc kubenswrapper[4715]: I1204 14:14:49.495167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.495358 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.495407 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:53.495390845 +0000 UTC m=+1050.564109050 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "webhook-server-cert" not found Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.495420 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:14:49 crc kubenswrapper[4715]: E1204 14:14:49.495522 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:14:53.495499018 +0000 UTC m=+1050.564217293 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "metrics-server-cert" not found Dec 04 14:14:50 crc kubenswrapper[4715]: E1204 14:14:50.447446 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" podUID="43170033-ddda-4fdf-b0e7-8d3b6d15e4f9" Dec 04 14:14:50 crc kubenswrapper[4715]: E1204 14:14:50.449440 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" podUID="88013762-6dd2-4994-b34e-21c89ed5afee" Dec 04 14:14:50 crc kubenswrapper[4715]: E1204 14:14:50.457642 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" podUID="d7fc7c62-f1ba-4fcc-9b82-b980988c90cf" Dec 04 14:14:50 crc kubenswrapper[4715]: E1204 14:14:50.457794 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" podUID="ec0c61ca-d902-4703-85bc-87b55736fe78" Dec 04 14:14:50 crc kubenswrapper[4715]: E1204 14:14:50.457883 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" podUID="4b1a2407-c585-4f22-95db-a2331a69623d" Dec 04 14:14:50 crc kubenswrapper[4715]: E1204 14:14:50.458001 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" podUID="4238708b-a19f-485f-a8a7-66b58aac168f" Dec 04 14:14:52 crc kubenswrapper[4715]: I1204 14:14:52.408422 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:14:52 crc kubenswrapper[4715]: E1204 14:14:52.408643 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:52 crc kubenswrapper[4715]: E1204 14:14:52.408942 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert podName:8a9a0373-b39e-496a-aba7-456bddcf3a65 nodeName:}" failed. No retries permitted until 2025-12-04 14:15:00.408901876 +0000 UTC m=+1057.477620151 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert") pod "infra-operator-controller-manager-57548d458d-f8cgr" (UID: "8a9a0373-b39e-496a-aba7-456bddcf3a65") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:14:52 crc kubenswrapper[4715]: I1204 14:14:52.798220 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:14:52 crc kubenswrapper[4715]: E1204 14:14:52.798408 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:52 crc kubenswrapper[4715]: E1204 14:14:52.798459 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert podName:73189a74-45fd-4ea8-92ce-bfa038531159 nodeName:}" failed. No retries permitted until 2025-12-04 14:15:00.798442422 +0000 UTC m=+1057.867160637 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" (UID: "73189a74-45fd-4ea8-92ce-bfa038531159") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:14:53 crc kubenswrapper[4715]: I1204 14:14:53.548593 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:53 crc kubenswrapper[4715]: I1204 14:14:53.549021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:14:53 crc kubenswrapper[4715]: E1204 14:14:53.548794 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:14:53 crc kubenswrapper[4715]: E1204 14:14:53.549217 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:15:01.549201929 +0000 UTC m=+1058.617920134 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "webhook-server-cert" not found Dec 04 14:14:53 crc kubenswrapper[4715]: E1204 14:14:53.549163 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:14:53 crc kubenswrapper[4715]: E1204 14:14:53.549528 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:15:01.549518837 +0000 UTC m=+1058.618237042 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "metrics-server-cert" not found Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.153671 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr"] Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.155482 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.157988 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.158720 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.162739 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr"] Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.429278 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.429341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46bcad15-7a3f-4bc0-8ae7-344881de90e5-secret-volume\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.429365 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46bcad15-7a3f-4bc0-8ae7-344881de90e5-config-volume\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.429390 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp8mb\" (UniqueName: \"kubernetes.io/projected/46bcad15-7a3f-4bc0-8ae7-344881de90e5-kube-api-access-lp8mb\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.447241 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8a9a0373-b39e-496a-aba7-456bddcf3a65-cert\") pod \"infra-operator-controller-manager-57548d458d-f8cgr\" (UID: \"8a9a0373-b39e-496a-aba7-456bddcf3a65\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.530782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46bcad15-7a3f-4bc0-8ae7-344881de90e5-secret-volume\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.530837 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46bcad15-7a3f-4bc0-8ae7-344881de90e5-config-volume\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.530873 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp8mb\" (UniqueName: \"kubernetes.io/projected/46bcad15-7a3f-4bc0-8ae7-344881de90e5-kube-api-access-lp8mb\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.532673 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46bcad15-7a3f-4bc0-8ae7-344881de90e5-config-volume\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.535382 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46bcad15-7a3f-4bc0-8ae7-344881de90e5-secret-volume\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.551299 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.555201 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp8mb\" (UniqueName: \"kubernetes.io/projected/46bcad15-7a3f-4bc0-8ae7-344881de90e5-kube-api-access-lp8mb\") pod \"collect-profiles-29414295-lw8cr\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.790448 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.880590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:15:00 crc kubenswrapper[4715]: I1204 14:15:00.890592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73189a74-45fd-4ea8-92ce-bfa038531159-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w\" (UID: \"73189a74-45fd-4ea8-92ce-bfa038531159\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:15:01 crc kubenswrapper[4715]: I1204 14:15:01.142861 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:15:01 crc kubenswrapper[4715]: I1204 14:15:01.590476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:01 crc kubenswrapper[4715]: E1204 14:15:01.590589 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:15:01 crc kubenswrapper[4715]: I1204 14:15:01.590597 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:01 crc kubenswrapper[4715]: E1204 14:15:01.590653 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:15:17.59063215 +0000 UTC m=+1074.659350375 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "webhook-server-cert" not found Dec 04 14:15:01 crc kubenswrapper[4715]: E1204 14:15:01.590693 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:15:01 crc kubenswrapper[4715]: E1204 14:15:01.590730 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs podName:00ef5230-a1c9-4609-ba71-e90760f29a15 nodeName:}" failed. No retries permitted until 2025-12-04 14:15:17.590719322 +0000 UTC m=+1074.659437537 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs") pod "openstack-operator-controller-manager-6b8c75fd85-txxk5" (UID: "00ef5230-a1c9-4609-ba71-e90760f29a15") : secret "metrics-server-cert" not found Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.831390 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.831966 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.832017 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.832699 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60dea1351c777fadc6f4bbc06b373dac4289c7f94189506529170a4446c48e46"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.832752 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://60dea1351c777fadc6f4bbc06b373dac4289c7f94189506529170a4446c48e46" gracePeriod=600 Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.965111 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="60dea1351c777fadc6f4bbc06b373dac4289c7f94189506529170a4446c48e46" exitCode=0 Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.965167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"60dea1351c777fadc6f4bbc06b373dac4289c7f94189506529170a4446c48e46"} Dec 04 14:15:08 crc kubenswrapper[4715]: I1204 14:15:08.965205 4715 scope.go:117] "RemoveContainer" containerID="e37f663cd84616cad5630c875a686e056c8f65437b236f02818772405f217b98" Dec 04 14:15:09 crc kubenswrapper[4715]: E1204 14:15:09.482459 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 04 14:15:09 crc kubenswrapper[4715]: E1204 14:15:09.482705 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p98ww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vnndg_openstack-operators(ab0e965e-1d56-4b31-8cf1-6a21892d38ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:10 crc kubenswrapper[4715]: E1204 14:15:10.656826 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 04 14:15:10 crc kubenswrapper[4715]: E1204 14:15:10.657333 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jl89v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-w2drg_openstack-operators(9d367397-9c5c-4ef7-bf95-91d0a453cf04): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:12 crc kubenswrapper[4715]: E1204 14:15:12.003311 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 04 14:15:12 crc kubenswrapper[4715]: E1204 14:15:12.003505 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hbjf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-xncpr_openstack-operators(5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:12 crc kubenswrapper[4715]: I1204 14:15:12.008621 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:15:13 crc kubenswrapper[4715]: E1204 14:15:13.704850 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 04 14:15:13 crc kubenswrapper[4715]: E1204 14:15:13.705687 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z4dg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-qq7m7_openstack-operators(6d5d27a6-eba8-4699-a415-e0e8c283d9cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:14 crc kubenswrapper[4715]: E1204 14:15:14.507363 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 04 14:15:14 crc kubenswrapper[4715]: E1204 14:15:14.507551 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vj9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-fdgb9_openstack-operators(6efc64e6-15fa-4c60-9a4e-18d326e4c759): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:15 crc kubenswrapper[4715]: E1204 14:15:15.333519 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 04 14:15:15 crc kubenswrapper[4715]: E1204 14:15:15.333695 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lk55c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-wwc57_openstack-operators(84876a00-902b-41f4-af69-f39d588e34fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:16 crc kubenswrapper[4715]: E1204 14:15:16.340483 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 04 14:15:16 crc kubenswrapper[4715]: E1204 14:15:16.340961 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c4k58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-5glkm_openstack-operators(81510632-c76c-4465-8dec-d8d9c3e8c88d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:17 crc kubenswrapper[4715]: E1204 14:15:17.520613 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 04 14:15:17 crc kubenswrapper[4715]: E1204 14:15:17.521051 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nnsxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-hhjwx_openstack-operators(45a4ad9a-e82a-466a-954b-12127626b577): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:17 crc kubenswrapper[4715]: I1204 14:15:17.613982 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:17 crc kubenswrapper[4715]: I1204 14:15:17.614114 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:17 crc kubenswrapper[4715]: I1204 14:15:17.621709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-metrics-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:17 crc kubenswrapper[4715]: I1204 14:15:17.637029 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00ef5230-a1c9-4609-ba71-e90760f29a15-webhook-certs\") pod \"openstack-operator-controller-manager-6b8c75fd85-txxk5\" (UID: \"00ef5230-a1c9-4609-ba71-e90760f29a15\") " pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:17 crc kubenswrapper[4715]: I1204 14:15:17.918164 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:18 crc kubenswrapper[4715]: E1204 14:15:18.177463 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 04 14:15:18 crc kubenswrapper[4715]: E1204 14:15:18.177660 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cgq9k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-grcg8_openstack-operators(96aa918a-abfb-44f6-bfa3-9849ef0f51e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:22 crc kubenswrapper[4715]: E1204 14:15:22.498203 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 04 14:15:22 crc kubenswrapper[4715]: E1204 14:15:22.498952 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4qdcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-5m69r_openstack-operators(4b1a2407-c585-4f22-95db-a2331a69623d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:24 crc kubenswrapper[4715]: E1204 14:15:24.548788 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 04 14:15:24 crc kubenswrapper[4715]: E1204 14:15:24.549314 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6kdqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-dg7m6_openstack-operators(4238708b-a19f-485f-a8a7-66b58aac168f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:25 crc kubenswrapper[4715]: E1204 14:15:25.151589 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 04 14:15:25 crc kubenswrapper[4715]: E1204 14:15:25.152076 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v6xnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-rtj2r_openstack-operators(9c67b2dc-8ac4-4dd0-af9f-19e55d368092): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:25 crc kubenswrapper[4715]: E1204 14:15:25.758236 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 04 14:15:25 crc kubenswrapper[4715]: E1204 14:15:25.758442 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7csxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-9g592_openstack-operators(43170033-ddda-4fdf-b0e7-8d3b6d15e4f9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:25 crc kubenswrapper[4715]: E1204 14:15:25.759590 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" podUID="43170033-ddda-4fdf-b0e7-8d3b6d15e4f9" Dec 04 14:15:26 crc kubenswrapper[4715]: I1204 14:15:26.546113 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w"] Dec 04 14:15:26 crc kubenswrapper[4715]: I1204 14:15:26.792192 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr"] Dec 04 14:15:26 crc kubenswrapper[4715]: I1204 14:15:26.882434 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr"] Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.066106 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5"] Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.257223 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"d601d4f16680a55d11b4648fd80ee11fd0f65f95595398da0f63f96f86a77a1f"} Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.260303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" event={"ID":"46bcad15-7a3f-4bc0-8ae7-344881de90e5","Type":"ContainerStarted","Data":"9c73b77cb28a515f5e17ea651ad8868a82bbfaecd94e793dbab964ae5c3e7bbd"} Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.262565 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" event={"ID":"fbc62f1d-2aa8-4b6b-8693-a16c349461d3","Type":"ContainerStarted","Data":"b71bc946b2e58c700c8cdea680d788ddab006aa7e03e08269ed54c16495ff207"} Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.264846 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" event={"ID":"3521ae52-a647-4190-8bbd-1100bb31dae5","Type":"ContainerStarted","Data":"0958181cdbdec01573de47da0fde78c16e41b7c00c14fc276273d331acddda34"} Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.267022 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" event={"ID":"cec0926c-ac5c-4ef2-aefc-a81c7885d810","Type":"ContainerStarted","Data":"620d3fe7739234648fa4386d160e08c3eda6e9a76497d03f25009d28c8386c2b"} Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.269238 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" event={"ID":"73189a74-45fd-4ea8-92ce-bfa038531159","Type":"ContainerStarted","Data":"758f01d7d15adf0f1dc46f3e2f70693adce41b819a5afde69eb191be0bea4f32"} Dec 04 14:15:27 crc kubenswrapper[4715]: I1204 14:15:27.283080 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" event={"ID":"8a9a0373-b39e-496a-aba7-456bddcf3a65","Type":"ContainerStarted","Data":"11fd6321258d8fe347466e81f040eb0e1f95583c2b448cb8ed9a32ab35c25c17"} Dec 04 14:15:28 crc kubenswrapper[4715]: I1204 14:15:28.318188 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" event={"ID":"00ef5230-a1c9-4609-ba71-e90760f29a15","Type":"ContainerStarted","Data":"abd103680619423c17184b8f1cd88be86f826c2ea6eae55bed2e00f54eb9408d"} Dec 04 14:15:29 crc kubenswrapper[4715]: I1204 14:15:29.324763 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" event={"ID":"ec0c61ca-d902-4703-85bc-87b55736fe78","Type":"ContainerStarted","Data":"bd8d98e7fc311b39136c085fa4eaff8a8b04a44f525ae7c1d08ea3e1abac611a"} Dec 04 14:15:30 crc kubenswrapper[4715]: I1204 14:15:30.366609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" event={"ID":"88013762-6dd2-4994-b34e-21c89ed5afee","Type":"ContainerStarted","Data":"c654dfa70a8ad0f90ec06140aeab742d69e5f2f3df02233b63bb175051da70f5"} Dec 04 14:15:36 crc kubenswrapper[4715]: E1204 14:15:36.991942 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 14:15:36 crc kubenswrapper[4715]: E1204 14:15:36.992696 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9wsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-8ztmn_openstack-operators(d58523b1-200c-4b54-ab7c-fab0ee9ff571): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:36 crc kubenswrapper[4715]: E1204 14:15:36.993947 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" podUID="d58523b1-200c-4b54-ab7c-fab0ee9ff571" Dec 04 14:15:37 crc kubenswrapper[4715]: E1204 14:15:37.708978 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 14:15:37 crc kubenswrapper[4715]: E1204 14:15:37.709474 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lk55c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-wwc57_openstack-operators(84876a00-902b-41f4-af69-f39d588e34fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:37 crc kubenswrapper[4715]: E1204 14:15:37.710830 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" podUID="84876a00-902b-41f4-af69-f39d588e34fe" Dec 04 14:15:39 crc kubenswrapper[4715]: E1204 14:15:39.182161 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" podUID="43170033-ddda-4fdf-b0e7-8d3b6d15e4f9" Dec 04 14:15:51 crc kubenswrapper[4715]: E1204 14:15:51.720017 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 04 14:15:51 crc kubenswrapper[4715]: E1204 14:15:51.721116 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wt6v6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w_openstack-operators(73189a74-45fd-4ea8-92ce-bfa038531159): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:51 crc kubenswrapper[4715]: I1204 14:15:51.936692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" event={"ID":"46bcad15-7a3f-4bc0-8ae7-344881de90e5","Type":"ContainerStarted","Data":"3cf61e5e6767afdc90f9ce2d205b45315fe6ad2449aee48f6790d497229b2482"} Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.206661 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" podStartSLOduration=52.206643156 podStartE2EDuration="52.206643156s" podCreationTimestamp="2025-12-04 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:51.958470476 +0000 UTC m=+1109.027188691" watchObservedRunningTime="2025-12-04 14:15:52.206643156 +0000 UTC m=+1109.275361371" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.442087 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.442779 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xll8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-gv5kg_openstack-operators(fbc62f1d-2aa8-4b6b-8693-a16c349461d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.445165 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" podUID="fbc62f1d-2aa8-4b6b-8693-a16c349461d3" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.446928 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.447245 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v6l6c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-f8cgr_openstack-operators(8a9a0373-b39e-496a-aba7-456bddcf3a65): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.447557 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.447723 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-djpbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-cw997_openstack-operators(ec0c61ca-d902-4703-85bc-87b55736fe78): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.449080 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" podUID="ec0c61ca-d902-4703-85bc-87b55736fe78" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.474430 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.474622 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dd25j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-sg77c_openstack-operators(88013762-6dd2-4994-b34e-21c89ed5afee): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 04 14:15:52 crc kubenswrapper[4715]: E1204 14:15:52.476651 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" podUID="88013762-6dd2-4994-b34e-21c89ed5afee" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.950718 4715 generic.go:334] "Generic (PLEG): container finished" podID="46bcad15-7a3f-4bc0-8ae7-344881de90e5" containerID="3cf61e5e6767afdc90f9ce2d205b45315fe6ad2449aee48f6790d497229b2482" exitCode=0 Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.950944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" event={"ID":"46bcad15-7a3f-4bc0-8ae7-344881de90e5","Type":"ContainerDied","Data":"3cf61e5e6767afdc90f9ce2d205b45315fe6ad2449aee48f6790d497229b2482"} Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.953236 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" event={"ID":"d7fc7c62-f1ba-4fcc-9b82-b980988c90cf","Type":"ContainerStarted","Data":"11d6f61746508ac86db83286c9db9fe57c6df0c44e7664d3cd9200c2b3fa5d29"} Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.956346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" event={"ID":"00ef5230-a1c9-4609-ba71-e90760f29a15","Type":"ContainerStarted","Data":"e33f7e32069b1fa3684167e28cd817408f76053ac31a308224ae0e8f20d12afd"} Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.957559 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.957600 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.958254 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.958505 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.962198 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.964389 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" Dec 04 14:15:52 crc kubenswrapper[4715]: I1204 14:15:52.964461 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" Dec 04 14:15:53 crc kubenswrapper[4715]: I1204 14:15:53.017219 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" podStartSLOduration=68.017194327 podStartE2EDuration="1m8.017194327s" podCreationTimestamp="2025-12-04 14:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:53.011423133 +0000 UTC m=+1110.080141368" watchObservedRunningTime="2025-12-04 14:15:53.017194327 +0000 UTC m=+1110.085912562" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.172586 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" podUID="6d5d27a6-eba8-4699-a415-e0e8c283d9cf" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.482422 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" podUID="4b1a2407-c585-4f22-95db-a2331a69623d" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.483914 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" podUID="5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.509567 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" podUID="96aa918a-abfb-44f6-bfa3-9849ef0f51e4" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.541291 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" podUID="ab0e965e-1d56-4b31-8cf1-6a21892d38ca" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.549985 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" podUID="4238708b-a19f-485f-a8a7-66b58aac168f" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.578421 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" podUID="73189a74-45fd-4ea8-92ce-bfa038531159" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.591886 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" podUID="45a4ad9a-e82a-466a-954b-12127626b577" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.601149 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" podUID="9d367397-9c5c-4ef7-bf95-91d0a453cf04" Dec 04 14:15:53 crc kubenswrapper[4715]: E1204 14:15:53.758558 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" podUID="81510632-c76c-4465-8dec-d8d9c3e8c88d" Dec 04 14:15:53 crc kubenswrapper[4715]: I1204 14:15:53.978948 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" event={"ID":"45a4ad9a-e82a-466a-954b-12127626b577","Type":"ContainerStarted","Data":"9dddc9f1d72897825c3e380807f67208931eb9756e23da7f6fe8c2c63625fb54"} Dec 04 14:15:53 crc kubenswrapper[4715]: I1204 14:15:53.998386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" event={"ID":"84876a00-902b-41f4-af69-f39d588e34fe","Type":"ContainerStarted","Data":"06dba1ef5816462d5dc7baed96d149ba011e392f5be25f2230dccc09d432f48c"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.008315 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" event={"ID":"3521ae52-a647-4190-8bbd-1100bb31dae5","Type":"ContainerStarted","Data":"00ea33dd1c0de9d810be8f80a17eca033d8b40b74e37ff2103dd12493c82105d"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.009352 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.013223 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" event={"ID":"81510632-c76c-4465-8dec-d8d9c3e8c88d","Type":"ContainerStarted","Data":"9f106bdc44cad8ce5f8e5c1a7ec0a3a7db46a1920132c352f9d7de811eddf4f9"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.014938 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.022587 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" event={"ID":"6d5d27a6-eba8-4699-a415-e0e8c283d9cf","Type":"ContainerStarted","Data":"c883427767fc06de1102abc8d7117f910b004749396adbe7cc845640ab856e57"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.031290 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" event={"ID":"9d367397-9c5c-4ef7-bf95-91d0a453cf04","Type":"ContainerStarted","Data":"be27b5fc5b78ae5f05f36cc909895f70a0bc03dfb0f8c525879edb562141f14e"} Dec 04 14:15:54 crc kubenswrapper[4715]: E1204 14:15:54.033114 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" podUID="9c67b2dc-8ac4-4dd0-af9f-19e55d368092" Dec 04 14:15:54 crc kubenswrapper[4715]: E1204 14:15:54.041761 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" podUID="6efc64e6-15fa-4c60-9a4e-18d326e4c759" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.059342 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" event={"ID":"43170033-ddda-4fdf-b0e7-8d3b6d15e4f9","Type":"ContainerStarted","Data":"3d228d7ee67e8f32ad3c2dc8fe5fbc34610e38379f662094995a302f776f8dc2"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.064649 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" event={"ID":"ab0e965e-1d56-4b31-8cf1-6a21892d38ca","Type":"ContainerStarted","Data":"7217dd1bde107894523b29c483a7190754a5cb0d4a20f04ac9466c37548efccf"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.077794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" event={"ID":"cec0926c-ac5c-4ef2-aefc-a81c7885d810","Type":"ContainerStarted","Data":"3b9740f77bab27037dec6156d449371047056a12cbd9b7cb902878dffbb6b70b"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.080068 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.088951 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.093222 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" event={"ID":"4b1a2407-c585-4f22-95db-a2331a69623d","Type":"ContainerStarted","Data":"e9cc9c4857bc6e41ef4b56166f2a5ac899ec3cc52fc3ba618de8fc7c88d65626"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.113413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" event={"ID":"73189a74-45fd-4ea8-92ce-bfa038531159","Type":"ContainerStarted","Data":"9c68a9974f7ea3d3c23f40d3037d3b5d054cc2c3891814570b27220a27908310"} Dec 04 14:15:54 crc kubenswrapper[4715]: E1204 14:15:54.170127 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" podUID="73189a74-45fd-4ea8-92ce-bfa038531159" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.182591 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6hbp2" podStartSLOduration=5.994178078 podStartE2EDuration="1m10.182572513s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.544113878 +0000 UTC m=+1045.612832093" lastFinishedPulling="2025-12-04 14:15:52.732508313 +0000 UTC m=+1109.801226528" observedRunningTime="2025-12-04 14:15:54.133580676 +0000 UTC m=+1111.202298891" watchObservedRunningTime="2025-12-04 14:15:54.182572513 +0000 UTC m=+1111.251290728" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.213096 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" event={"ID":"96aa918a-abfb-44f6-bfa3-9849ef0f51e4","Type":"ContainerStarted","Data":"4ca71ca20e21e6233a3327bd9484c8ff27b8b1fe0b105648aa9b97567b34d20f"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.496788 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" event={"ID":"d7fc7c62-f1ba-4fcc-9b82-b980988c90cf","Type":"ContainerStarted","Data":"2e85e48d8aebb22b3f886824e6144c3da513225e3425fa267de03fcb97922ae9"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.497761 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.528974 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9g592" podStartSLOduration=5.54972525 podStartE2EDuration="1m9.528956333s" podCreationTimestamp="2025-12-04 14:14:45 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.750146026 +0000 UTC m=+1045.818864241" lastFinishedPulling="2025-12-04 14:15:52.729377109 +0000 UTC m=+1109.798095324" observedRunningTime="2025-12-04 14:15:54.463503277 +0000 UTC m=+1111.532221492" watchObservedRunningTime="2025-12-04 14:15:54.528956333 +0000 UTC m=+1111.597674548" Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.539252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" event={"ID":"4238708b-a19f-485f-a8a7-66b58aac168f","Type":"ContainerStarted","Data":"38ffd1d12d2a920d20c29315cfb3c16ec87f0164cea3e6e8d56fff77a7d54494"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.567943 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" event={"ID":"5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd","Type":"ContainerStarted","Data":"aca7146af0633abe6901b9123ebf7907b421da60670842cf3169e8003a880df1"} Dec 04 14:15:54 crc kubenswrapper[4715]: I1204 14:15:54.832593 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-xc2z5" podStartSLOduration=6.835116365 podStartE2EDuration="1m10.832570601s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.659175853 +0000 UTC m=+1045.727894078" lastFinishedPulling="2025-12-04 14:15:52.656630099 +0000 UTC m=+1109.725348314" observedRunningTime="2025-12-04 14:15:54.831288406 +0000 UTC m=+1111.900006641" watchObservedRunningTime="2025-12-04 14:15:54.832570601 +0000 UTC m=+1111.901288836" Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.107886 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" podStartSLOduration=33.707724606 podStartE2EDuration="1m11.107864484s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.735882086 +0000 UTC m=+1045.804600291" lastFinishedPulling="2025-12-04 14:15:26.136021954 +0000 UTC m=+1083.204740169" observedRunningTime="2025-12-04 14:15:55.101601947 +0000 UTC m=+1112.170320162" watchObservedRunningTime="2025-12-04 14:15:55.107864484 +0000 UTC m=+1112.176582699" Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.605439 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" event={"ID":"84876a00-902b-41f4-af69-f39d588e34fe","Type":"ContainerStarted","Data":"c8a3ee8be40e0096273ac34226f7918e2bc0c01734e3dc8352da6e1c6b2a9fcc"} Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.606714 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.623874 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" event={"ID":"fbc62f1d-2aa8-4b6b-8693-a16c349461d3","Type":"ContainerStarted","Data":"724fabecc6ce0b3cca9fc92418463fb07d802931c0f85bb663b838aa3b6863f4"} Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.630416 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" event={"ID":"9c67b2dc-8ac4-4dd0-af9f-19e55d368092","Type":"ContainerStarted","Data":"14d12040dc248eeb281226a2d4ce76918fb99c026f56aa14727533b77f01713d"} Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.631963 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" podStartSLOduration=7.636557513 podStartE2EDuration="1m11.631949184s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.650654266 +0000 UTC m=+1045.719372481" lastFinishedPulling="2025-12-04 14:15:52.646045937 +0000 UTC m=+1109.714764152" observedRunningTime="2025-12-04 14:15:55.627851235 +0000 UTC m=+1112.696569460" watchObservedRunningTime="2025-12-04 14:15:55.631949184 +0000 UTC m=+1112.700667399" Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.637276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" event={"ID":"88013762-6dd2-4994-b34e-21c89ed5afee","Type":"ContainerStarted","Data":"409f9617df0510ec20a4bcb01d80e2bcaf1c8dbd2e05ae61fd77ecbfec8a6524"} Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.644529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" event={"ID":"6efc64e6-15fa-4c60-9a4e-18d326e4c759","Type":"ContainerStarted","Data":"303583ff8eb631f5e37a2af3cd9d09468866c9cf79fe5b027879ebe8574b8eaf"} Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.648368 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gv5kg" podStartSLOduration=40.157474716 podStartE2EDuration="1m11.648345331s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:46.780363491 +0000 UTC m=+1043.849081706" lastFinishedPulling="2025-12-04 14:15:18.271234106 +0000 UTC m=+1075.339952321" observedRunningTime="2025-12-04 14:15:55.645787443 +0000 UTC m=+1112.714505658" watchObservedRunningTime="2025-12-04 14:15:55.648345331 +0000 UTC m=+1112.717063546" Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.661418 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" event={"ID":"ec0c61ca-d902-4703-85bc-87b55736fe78","Type":"ContainerStarted","Data":"8b3bd6ff4f0359f70332b4065b653a953dfbbe3e1026c5926ce6bbe5497ae005"} Dec 04 14:15:55 crc kubenswrapper[4715]: E1204 14:15:55.663719 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" podUID="73189a74-45fd-4ea8-92ce-bfa038531159" Dec 04 14:15:55 crc kubenswrapper[4715]: I1204 14:15:55.724062 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-sg77c" podStartSLOduration=34.744533436 podStartE2EDuration="1m11.7240241s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.765423733 +0000 UTC m=+1045.834141948" lastFinishedPulling="2025-12-04 14:15:25.744914397 +0000 UTC m=+1082.813632612" observedRunningTime="2025-12-04 14:15:55.722183061 +0000 UTC m=+1112.790901276" watchObservedRunningTime="2025-12-04 14:15:55.7240241 +0000 UTC m=+1112.792742315" Dec 04 14:15:56 crc kubenswrapper[4715]: I1204 14:15:56.980734 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.021596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46bcad15-7a3f-4bc0-8ae7-344881de90e5-secret-volume\") pod \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.021598 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-cw997" podStartSLOduration=36.029340088 podStartE2EDuration="1m13.021576712s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.753506915 +0000 UTC m=+1045.822225130" lastFinishedPulling="2025-12-04 14:15:25.745743539 +0000 UTC m=+1082.814461754" observedRunningTime="2025-12-04 14:15:55.883477353 +0000 UTC m=+1112.952195568" watchObservedRunningTime="2025-12-04 14:15:57.021576712 +0000 UTC m=+1114.090294927" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.021667 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46bcad15-7a3f-4bc0-8ae7-344881de90e5-config-volume\") pod \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.021711 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp8mb\" (UniqueName: \"kubernetes.io/projected/46bcad15-7a3f-4bc0-8ae7-344881de90e5-kube-api-access-lp8mb\") pod \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\" (UID: \"46bcad15-7a3f-4bc0-8ae7-344881de90e5\") " Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.090679 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46bcad15-7a3f-4bc0-8ae7-344881de90e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "46bcad15-7a3f-4bc0-8ae7-344881de90e5" (UID: "46bcad15-7a3f-4bc0-8ae7-344881de90e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.095827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46bcad15-7a3f-4bc0-8ae7-344881de90e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "46bcad15-7a3f-4bc0-8ae7-344881de90e5" (UID: "46bcad15-7a3f-4bc0-8ae7-344881de90e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.108466 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46bcad15-7a3f-4bc0-8ae7-344881de90e5-kube-api-access-lp8mb" (OuterVolumeSpecName: "kube-api-access-lp8mb") pod "46bcad15-7a3f-4bc0-8ae7-344881de90e5" (UID: "46bcad15-7a3f-4bc0-8ae7-344881de90e5"). InnerVolumeSpecName "kube-api-access-lp8mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.123001 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46bcad15-7a3f-4bc0-8ae7-344881de90e5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.123060 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46bcad15-7a3f-4bc0-8ae7-344881de90e5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.123078 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp8mb\" (UniqueName: \"kubernetes.io/projected/46bcad15-7a3f-4bc0-8ae7-344881de90e5-kube-api-access-lp8mb\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:57 crc kubenswrapper[4715]: E1204 14:15:57.127862 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" podUID="8a9a0373-b39e-496a-aba7-456bddcf3a65" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.680740 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" event={"ID":"d58523b1-200c-4b54-ab7c-fab0ee9ff571","Type":"ContainerStarted","Data":"45b94f8e2d00f6cac5646ccacbbc2475ae16125b5c3fa94cba510ae45ef14c04"} Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.681967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" event={"ID":"8a9a0373-b39e-496a-aba7-456bddcf3a65","Type":"ContainerStarted","Data":"6681a34addb63856b5a8d0c7fddd38e6145f557e2c880945db79133e994f77c7"} Dec 04 14:15:57 crc kubenswrapper[4715]: E1204 14:15:57.685438 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" podUID="8a9a0373-b39e-496a-aba7-456bddcf3a65" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.689528 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" event={"ID":"45a4ad9a-e82a-466a-954b-12127626b577","Type":"ContainerStarted","Data":"020664eb6b3b6be6ec35e78c91b261271b0539ad326752586cbef3b12799d72e"} Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.690245 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.692884 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.693354 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr" event={"ID":"46bcad15-7a3f-4bc0-8ae7-344881de90e5","Type":"ContainerDied","Data":"9c73b77cb28a515f5e17ea651ad8868a82bbfaecd94e793dbab964ae5c3e7bbd"} Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.693379 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c73b77cb28a515f5e17ea651ad8868a82bbfaecd94e793dbab964ae5c3e7bbd" Dec 04 14:15:57 crc kubenswrapper[4715]: I1204 14:15:57.852763 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" podStartSLOduration=5.195837948 podStartE2EDuration="1m13.852743833s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.654239821 +0000 UTC m=+1045.722958036" lastFinishedPulling="2025-12-04 14:15:57.311145706 +0000 UTC m=+1114.379863921" observedRunningTime="2025-12-04 14:15:57.851658904 +0000 UTC m=+1114.920377109" watchObservedRunningTime="2025-12-04 14:15:57.852743833 +0000 UTC m=+1114.921462058" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.271352 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6b8c75fd85-txxk5" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.701562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" event={"ID":"ab0e965e-1d56-4b31-8cf1-6a21892d38ca","Type":"ContainerStarted","Data":"fa662e4e43144f851f5003f1bab6fedbdb4d85f740c4df13e919b3afb9200e8a"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.702766 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.705078 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" event={"ID":"4b1a2407-c585-4f22-95db-a2331a69623d","Type":"ContainerStarted","Data":"9d7e3e8c14f223c6bcb5f40235c93e9c75bad26e985b671162319f49e23951e8"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.705749 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.707737 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" event={"ID":"6d5d27a6-eba8-4699-a415-e0e8c283d9cf","Type":"ContainerStarted","Data":"e3ccdc3b3ae74d32c13c058df282307fd871c323f2f1f421c1cabda12315d8e7"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.708279 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.710659 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" event={"ID":"6efc64e6-15fa-4c60-9a4e-18d326e4c759","Type":"ContainerStarted","Data":"7bd025039173bdc5d5c6a99adb7ac935d2d2087edc5f7298f7a549aa62f9ec4a"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.711348 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.713665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" event={"ID":"9d367397-9c5c-4ef7-bf95-91d0a453cf04","Type":"ContainerStarted","Data":"e643a793801ac20d4ba1341762da23b52b6ffa1aca7606ce19a9a01e0d8c4009"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.714273 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.716443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" event={"ID":"5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd","Type":"ContainerStarted","Data":"e8fd5dd4d71c5b2d9f658a703df959134fe22f8ca2be3e498655e16fe7de5a3e"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.716968 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.718920 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" event={"ID":"9c67b2dc-8ac4-4dd0-af9f-19e55d368092","Type":"ContainerStarted","Data":"7e7db99aab59d82ca7b5a72ca41388b519bbcbc7e6fce38698498629c3ea6549"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.719492 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.723817 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" event={"ID":"96aa918a-abfb-44f6-bfa3-9849ef0f51e4","Type":"ContainerStarted","Data":"fd85fd104eab0aca1565a5986446ae3cd7bfea4104f8d1671256c2540c5ade63"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.723946 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.726269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" event={"ID":"81510632-c76c-4465-8dec-d8d9c3e8c88d","Type":"ContainerStarted","Data":"9d82e250901064dcf87a13e95b25481e9df4ef789df106c7dac1efc0b65abb1d"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.726864 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.729271 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" event={"ID":"4238708b-a19f-485f-a8a7-66b58aac168f","Type":"ContainerStarted","Data":"f6830274e322ce9468ffe1495a2dd49b80a6fa284c1cc346c4eb5f284cd06b78"} Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.729772 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:15:58 crc kubenswrapper[4715]: E1204 14:15:58.731343 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" podUID="8a9a0373-b39e-496a-aba7-456bddcf3a65" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.734512 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" podStartSLOduration=5.848298445 podStartE2EDuration="1m14.734492733s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.559141068 +0000 UTC m=+1045.627859283" lastFinishedPulling="2025-12-04 14:15:57.445335346 +0000 UTC m=+1114.514053571" observedRunningTime="2025-12-04 14:15:58.730738963 +0000 UTC m=+1115.799457188" watchObservedRunningTime="2025-12-04 14:15:58.734492733 +0000 UTC m=+1115.803210958" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.801451 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" podStartSLOduration=6.260017879 podStartE2EDuration="1m14.801428828s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.780220627 +0000 UTC m=+1045.848938842" lastFinishedPulling="2025-12-04 14:15:57.321631576 +0000 UTC m=+1114.390349791" observedRunningTime="2025-12-04 14:15:58.794157614 +0000 UTC m=+1115.862875839" watchObservedRunningTime="2025-12-04 14:15:58.801428828 +0000 UTC m=+1115.870147043" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.821835 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" podStartSLOduration=6.090170785 podStartE2EDuration="1m14.821816972s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.593839112 +0000 UTC m=+1045.662557327" lastFinishedPulling="2025-12-04 14:15:57.325485299 +0000 UTC m=+1114.394203514" observedRunningTime="2025-12-04 14:15:58.81761804 +0000 UTC m=+1115.886336255" watchObservedRunningTime="2025-12-04 14:15:58.821816972 +0000 UTC m=+1115.890535177" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.848193 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" podStartSLOduration=5.069937002 podStartE2EDuration="1m14.848161855s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:47.546619939 +0000 UTC m=+1044.615338154" lastFinishedPulling="2025-12-04 14:15:57.324844792 +0000 UTC m=+1114.393563007" observedRunningTime="2025-12-04 14:15:58.841118257 +0000 UTC m=+1115.909836472" watchObservedRunningTime="2025-12-04 14:15:58.848161855 +0000 UTC m=+1115.916880070" Dec 04 14:15:58 crc kubenswrapper[4715]: I1204 14:15:58.892272 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" podStartSLOduration=6.227415335 podStartE2EDuration="1m14.892246881s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.650328927 +0000 UTC m=+1045.719047152" lastFinishedPulling="2025-12-04 14:15:57.315160473 +0000 UTC m=+1114.383878698" observedRunningTime="2025-12-04 14:15:58.883649601 +0000 UTC m=+1115.952367816" watchObservedRunningTime="2025-12-04 14:15:58.892246881 +0000 UTC m=+1115.960965096" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.094386 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" podStartSLOduration=6.328158296 podStartE2EDuration="1m15.094355822s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.544169949 +0000 UTC m=+1045.612888154" lastFinishedPulling="2025-12-04 14:15:57.310367465 +0000 UTC m=+1114.379085680" observedRunningTime="2025-12-04 14:15:58.929213687 +0000 UTC m=+1115.997931922" watchObservedRunningTime="2025-12-04 14:15:59.094355822 +0000 UTC m=+1116.163074037" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.096849 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" podStartSLOduration=4.51338282 podStartE2EDuration="1m15.096834488s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:46.738956688 +0000 UTC m=+1043.807674903" lastFinishedPulling="2025-12-04 14:15:57.322408356 +0000 UTC m=+1114.391126571" observedRunningTime="2025-12-04 14:15:59.091564097 +0000 UTC m=+1116.160282332" watchObservedRunningTime="2025-12-04 14:15:59.096834488 +0000 UTC m=+1116.165552703" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.208579 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" podStartSLOduration=6.597523145 podStartE2EDuration="1m15.208554708s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.71050004 +0000 UTC m=+1045.779218255" lastFinishedPulling="2025-12-04 14:15:57.321531603 +0000 UTC m=+1114.390249818" observedRunningTime="2025-12-04 14:15:59.204255863 +0000 UTC m=+1116.272974088" watchObservedRunningTime="2025-12-04 14:15:59.208554708 +0000 UTC m=+1116.277272923" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.279916 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" podStartSLOduration=6.554330857 podStartE2EDuration="1m15.279893251s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.596435972 +0000 UTC m=+1045.665154187" lastFinishedPulling="2025-12-04 14:15:57.321998366 +0000 UTC m=+1114.390716581" observedRunningTime="2025-12-04 14:15:59.277133727 +0000 UTC m=+1116.345851942" watchObservedRunningTime="2025-12-04 14:15:59.279893251 +0000 UTC m=+1116.348611476" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.346727 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" podStartSLOduration=6.663308562 podStartE2EDuration="1m15.346704823s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.627877639 +0000 UTC m=+1045.696595854" lastFinishedPulling="2025-12-04 14:15:57.3112739 +0000 UTC m=+1114.379992115" observedRunningTime="2025-12-04 14:15:59.342207473 +0000 UTC m=+1116.410925698" watchObservedRunningTime="2025-12-04 14:15:59.346704823 +0000 UTC m=+1116.415423038" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.783095 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" event={"ID":"d58523b1-200c-4b54-ab7c-fab0ee9ff571","Type":"ContainerStarted","Data":"463fc59008eea6d8c067c3f7a2c8386a057cba0bf65130013deb7248b0193649"} Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.787597 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:15:59 crc kubenswrapper[4715]: I1204 14:15:59.813258 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" podStartSLOduration=7.238986824 podStartE2EDuration="1m15.813236998s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:14:48.73567575 +0000 UTC m=+1045.804393965" lastFinishedPulling="2025-12-04 14:15:57.309925924 +0000 UTC m=+1114.378644139" observedRunningTime="2025-12-04 14:15:59.805052719 +0000 UTC m=+1116.873770934" watchObservedRunningTime="2025-12-04 14:15:59.813236998 +0000 UTC m=+1116.881955213" Dec 04 14:16:02 crc kubenswrapper[4715]: I1204 14:16:02.205419 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8ztmn" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.617367 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-grcg8" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.799882 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-xncpr" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.842552 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hhjwx" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.864804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-snfrf" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.869406 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fdgb9" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.891577 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-5m69r" Dec 04 14:16:04 crc kubenswrapper[4715]: I1204 14:16:04.979965 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5glkm" Dec 04 14:16:05 crc kubenswrapper[4715]: I1204 14:16:05.006398 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-rtj2r" Dec 04 14:16:05 crc kubenswrapper[4715]: I1204 14:16:05.092407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dg7m6" Dec 04 14:16:05 crc kubenswrapper[4715]: I1204 14:16:05.231989 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-wwc57" Dec 04 14:16:05 crc kubenswrapper[4715]: I1204 14:16:05.233945 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vnndg" Dec 04 14:16:05 crc kubenswrapper[4715]: I1204 14:16:05.379220 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qq7m7" Dec 04 14:16:05 crc kubenswrapper[4715]: I1204 14:16:05.447668 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-w2drg" Dec 04 14:16:09 crc kubenswrapper[4715]: I1204 14:16:09.312051 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" event={"ID":"73189a74-45fd-4ea8-92ce-bfa038531159","Type":"ContainerStarted","Data":"7ae6881566e5a1febff4ce9f2e2dcdac956222dbded3cb5ee7e74f81efff7100"} Dec 04 14:16:09 crc kubenswrapper[4715]: I1204 14:16:09.312775 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:16:12 crc kubenswrapper[4715]: I1204 14:16:12.391923 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" event={"ID":"8a9a0373-b39e-496a-aba7-456bddcf3a65","Type":"ContainerStarted","Data":"d25bcc2193c2f179474dc5c924cdb5ee0f7539f0bf4ba03ecbac80c1d852e4cd"} Dec 04 14:16:12 crc kubenswrapper[4715]: I1204 14:16:12.392747 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:16:12 crc kubenswrapper[4715]: I1204 14:16:12.410759 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" podStartSLOduration=46.31772899 podStartE2EDuration="1m28.41073843s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:15:26.588230158 +0000 UTC m=+1083.656948373" lastFinishedPulling="2025-12-04 14:16:08.681239598 +0000 UTC m=+1125.749957813" observedRunningTime="2025-12-04 14:16:09.34827973 +0000 UTC m=+1126.416997955" watchObservedRunningTime="2025-12-04 14:16:12.41073843 +0000 UTC m=+1129.479456645" Dec 04 14:16:20 crc kubenswrapper[4715]: I1204 14:16:20.559922 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" Dec 04 14:16:20 crc kubenswrapper[4715]: I1204 14:16:20.583942 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-f8cgr" podStartSLOduration=51.840148612 podStartE2EDuration="1m36.583915716s" podCreationTimestamp="2025-12-04 14:14:44 +0000 UTC" firstStartedPulling="2025-12-04 14:15:27.021525939 +0000 UTC m=+1084.090244154" lastFinishedPulling="2025-12-04 14:16:11.765293043 +0000 UTC m=+1128.834011258" observedRunningTime="2025-12-04 14:16:12.414444359 +0000 UTC m=+1129.483162574" watchObservedRunningTime="2025-12-04 14:16:20.583915716 +0000 UTC m=+1137.652633941" Dec 04 14:16:21 crc kubenswrapper[4715]: I1204 14:16:21.149112 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.482705 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-stnmb"] Dec 04 14:16:34 crc kubenswrapper[4715]: E1204 14:16:34.483757 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bcad15-7a3f-4bc0-8ae7-344881de90e5" containerName="collect-profiles" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.483778 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bcad15-7a3f-4bc0-8ae7-344881de90e5" containerName="collect-profiles" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.483972 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bcad15-7a3f-4bc0-8ae7-344881de90e5" containerName="collect-profiles" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.485006 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.492594 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.492680 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.493475 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.507242 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-9s7gb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.508775 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-stnmb"] Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.575377 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s9hhd"] Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.576907 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.579011 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.605190 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s9hhd"] Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.658271 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03459255-c53c-4f8e-a1e3-b1a849a21f85-config\") pod \"dnsmasq-dns-675f4bcbfc-stnmb\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.658364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gc97\" (UniqueName: \"kubernetes.io/projected/03459255-c53c-4f8e-a1e3-b1a849a21f85-kube-api-access-9gc97\") pod \"dnsmasq-dns-675f4bcbfc-stnmb\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.759337 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03459255-c53c-4f8e-a1e3-b1a849a21f85-config\") pod \"dnsmasq-dns-675f4bcbfc-stnmb\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.759432 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbv2q\" (UniqueName: \"kubernetes.io/projected/e4cf7297-95ff-4dc0-ac99-528ff7225438-kube-api-access-xbv2q\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.759466 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.759489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gc97\" (UniqueName: \"kubernetes.io/projected/03459255-c53c-4f8e-a1e3-b1a849a21f85-kube-api-access-9gc97\") pod \"dnsmasq-dns-675f4bcbfc-stnmb\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.759510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-config\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.760869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03459255-c53c-4f8e-a1e3-b1a849a21f85-config\") pod \"dnsmasq-dns-675f4bcbfc-stnmb\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.786329 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gc97\" (UniqueName: \"kubernetes.io/projected/03459255-c53c-4f8e-a1e3-b1a849a21f85-kube-api-access-9gc97\") pod \"dnsmasq-dns-675f4bcbfc-stnmb\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.803993 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.862249 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.862307 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-config\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.862424 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbv2q\" (UniqueName: \"kubernetes.io/projected/e4cf7297-95ff-4dc0-ac99-528ff7225438-kube-api-access-xbv2q\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.863734 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.863799 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-config\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.881623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbv2q\" (UniqueName: \"kubernetes.io/projected/e4cf7297-95ff-4dc0-ac99-528ff7225438-kube-api-access-xbv2q\") pod \"dnsmasq-dns-78dd6ddcc-s9hhd\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:34 crc kubenswrapper[4715]: I1204 14:16:34.893991 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:16:35 crc kubenswrapper[4715]: I1204 14:16:35.129538 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-stnmb"] Dec 04 14:16:35 crc kubenswrapper[4715]: I1204 14:16:35.462617 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s9hhd"] Dec 04 14:16:35 crc kubenswrapper[4715]: I1204 14:16:35.550759 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" event={"ID":"e4cf7297-95ff-4dc0-ac99-528ff7225438","Type":"ContainerStarted","Data":"21280a0ed0181326a2595304151ec3d3b10c6663953d9ce62383e1686e98a39c"} Dec 04 14:16:35 crc kubenswrapper[4715]: I1204 14:16:35.552049 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" event={"ID":"03459255-c53c-4f8e-a1e3-b1a849a21f85","Type":"ContainerStarted","Data":"e616980656c160cfa10e3f1397f90a89a37d9e1f9299d3c7816b3269f6761e4d"} Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.535111 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-stnmb"] Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.564574 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-66hsn"] Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.567183 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.579847 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-66hsn"] Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.639900 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtmwd\" (UniqueName: \"kubernetes.io/projected/4ddf5478-4875-4285-98af-0739c4636797-kube-api-access-gtmwd\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.639950 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-config\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.640021 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.742539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtmwd\" (UniqueName: \"kubernetes.io/projected/4ddf5478-4875-4285-98af-0739c4636797-kube-api-access-gtmwd\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.742592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-config\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.742662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.743728 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.745016 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-config\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.779199 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtmwd\" (UniqueName: \"kubernetes.io/projected/4ddf5478-4875-4285-98af-0739c4636797-kube-api-access-gtmwd\") pod \"dnsmasq-dns-5ccc8479f9-66hsn\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.864761 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s9hhd"] Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.889598 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.905267 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5c45t"] Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.908429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:37 crc kubenswrapper[4715]: I1204 14:16:37.922155 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5c45t"] Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.049482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-config\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.049884 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.049913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgdww\" (UniqueName: \"kubernetes.io/projected/6d6533ec-6214-4aa6-828c-707f1d9f8f11-kube-api-access-kgdww\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.151277 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.151332 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgdww\" (UniqueName: \"kubernetes.io/projected/6d6533ec-6214-4aa6-828c-707f1d9f8f11-kube-api-access-kgdww\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.151383 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-config\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.152302 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-config\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.152868 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.224909 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgdww\" (UniqueName: \"kubernetes.io/projected/6d6533ec-6214-4aa6-828c-707f1d9f8f11-kube-api-access-kgdww\") pod \"dnsmasq-dns-57d769cc4f-5c45t\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.235005 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.442753 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-66hsn"] Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.615563 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" event={"ID":"4ddf5478-4875-4285-98af-0739c4636797","Type":"ContainerStarted","Data":"ccfaaf41cccdd5e5ac1e94651b2d0c359967d4c3c033be38f26c85345fb1be05"} Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.711048 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.712401 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.716713 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.716915 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.716946 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.717171 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.717317 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.717557 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-s8rcz" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.720578 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.727261 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771580 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771635 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771676 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771763 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771799 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrt59\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-kube-api-access-zrt59\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.771930 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.820595 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5c45t"] Dec 04 14:16:38 crc kubenswrapper[4715]: W1204 14:16:38.831202 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d6533ec_6214_4aa6_828c_707f1d9f8f11.slice/crio-3dd80161d389dded94aa3fa48c59e54a87f47d688d8e908a17ecf85c3c282de5 WatchSource:0}: Error finding container 3dd80161d389dded94aa3fa48c59e54a87f47d688d8e908a17ecf85c3c282de5: Status 404 returned error can't find the container with id 3dd80161d389dded94aa3fa48c59e54a87f47d688d8e908a17ecf85c3c282de5 Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873157 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873201 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrt59\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-kube-api-access-zrt59\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873220 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873282 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873301 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873326 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.873415 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.874899 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.874952 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.875224 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.875510 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.876249 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.876604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.879540 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.880042 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.880433 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.880641 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.899299 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrt59\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-kube-api-access-zrt59\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:38 crc kubenswrapper[4715]: I1204 14:16:38.904561 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.027439 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.028803 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.030911 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.031161 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.031396 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.031517 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.031635 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.031843 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tcnqh" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.032462 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.051906 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.065788 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.075830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.075901 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.075932 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-config-data\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.075961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.075990 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.076016 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/440f7f8d-a89d-4d6d-8280-c56fede1c91c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.076084 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.076121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.076152 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.076172 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/440f7f8d-a89d-4d6d-8280-c56fede1c91c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.076194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4rrs\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-kube-api-access-m4rrs\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.176969 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177047 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177074 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/440f7f8d-a89d-4d6d-8280-c56fede1c91c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177120 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4rrs\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-kube-api-access-m4rrs\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177195 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177223 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-config-data\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/440f7f8d-a89d-4d6d-8280-c56fede1c91c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.177871 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.178741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.179301 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.179613 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.180073 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.181517 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-config-data\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.182073 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/440f7f8d-a89d-4d6d-8280-c56fede1c91c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.182190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.184325 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.187410 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/440f7f8d-a89d-4d6d-8280-c56fede1c91c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.250499 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4rrs\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-kube-api-access-m4rrs\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.257102 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.375144 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.589704 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.628317 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" event={"ID":"6d6533ec-6214-4aa6-828c-707f1d9f8f11","Type":"ContainerStarted","Data":"3dd80161d389dded94aa3fa48c59e54a87f47d688d8e908a17ecf85c3c282de5"} Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.629809 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34885f9-cf98-4b1a-bbf3-c7af62f5f273","Type":"ContainerStarted","Data":"6563d11eed8877e6b691cd75b8a8b74020e65a4f46f79710abdbb14d6133e313"} Dec 04 14:16:39 crc kubenswrapper[4715]: I1204 14:16:39.940801 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:16:39 crc kubenswrapper[4715]: W1204 14:16:39.973404 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod440f7f8d_a89d_4d6d_8280_c56fede1c91c.slice/crio-c7ba1830e98359f3caa67efc3663c83e3ad2d8306c0af40e367d1ab878e8f775 WatchSource:0}: Error finding container c7ba1830e98359f3caa67efc3663c83e3ad2d8306c0af40e367d1ab878e8f775: Status 404 returned error can't find the container with id c7ba1830e98359f3caa67efc3663c83e3ad2d8306c0af40e367d1ab878e8f775 Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.588560 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.594233 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.594372 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.603437 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.603523 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.603729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xggpc" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.603933 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.612754 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.640281 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"440f7f8d-a89d-4d6d-8280-c56fede1c91c","Type":"ContainerStarted","Data":"c7ba1830e98359f3caa67efc3663c83e3ad2d8306c0af40e367d1ab878e8f775"} Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.704765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.704840 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d666280-e175-4f31-8e11-7af2d5e27e5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.704886 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7d666280-e175-4f31-8e11-7af2d5e27e5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.704918 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.704967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d666280-e175-4f31-8e11-7af2d5e27e5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.704995 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.705027 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll48h\" (UniqueName: \"kubernetes.io/projected/7d666280-e175-4f31-8e11-7af2d5e27e5a-kube-api-access-ll48h\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.705081 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806553 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d666280-e175-4f31-8e11-7af2d5e27e5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806657 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll48h\" (UniqueName: \"kubernetes.io/projected/7d666280-e175-4f31-8e11-7af2d5e27e5a-kube-api-access-ll48h\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806689 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d666280-e175-4f31-8e11-7af2d5e27e5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806816 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7d666280-e175-4f31-8e11-7af2d5e27e5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.806851 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.807987 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.808813 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.809804 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7d666280-e175-4f31-8e11-7af2d5e27e5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.810414 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.811290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d666280-e175-4f31-8e11-7af2d5e27e5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.817787 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d666280-e175-4f31-8e11-7af2d5e27e5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.817871 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d666280-e175-4f31-8e11-7af2d5e27e5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.833140 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll48h\" (UniqueName: \"kubernetes.io/projected/7d666280-e175-4f31-8e11-7af2d5e27e5a-kube-api-access-ll48h\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.835794 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7d666280-e175-4f31-8e11-7af2d5e27e5a\") " pod="openstack/openstack-galera-0" Dec 04 14:16:40 crc kubenswrapper[4715]: I1204 14:16:40.925961 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.033404 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.034736 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.043312 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.043443 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-j76rr" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.043653 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.043838 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.064756 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.126863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.126985 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e478a0b5-cea1-4639-a645-9660524ff9bd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.127028 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.127077 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7fr8\" (UniqueName: \"kubernetes.io/projected/e478a0b5-cea1-4639-a645-9660524ff9bd-kube-api-access-p7fr8\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.127098 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e478a0b5-cea1-4639-a645-9660524ff9bd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.127133 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.127155 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e478a0b5-cea1-4639-a645-9660524ff9bd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.127262 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.136665 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.137940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.148225 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.148546 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.148607 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-v5swg" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.155211 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkmsb\" (UniqueName: \"kubernetes.io/projected/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-kube-api-access-xkmsb\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228484 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-config-data\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7fr8\" (UniqueName: \"kubernetes.io/projected/e478a0b5-cea1-4639-a645-9660524ff9bd-kube-api-access-p7fr8\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e478a0b5-cea1-4639-a645-9660524ff9bd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228582 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228597 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e478a0b5-cea1-4639-a645-9660524ff9bd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-combined-ca-bundle\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228634 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-memcached-tls-certs\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228731 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-kolla-config\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.228798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e478a0b5-cea1-4639-a645-9660524ff9bd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.229181 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e478a0b5-cea1-4639-a645-9660524ff9bd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.229447 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.230609 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.230946 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.231334 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e478a0b5-cea1-4639-a645-9660524ff9bd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.234314 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e478a0b5-cea1-4639-a645-9660524ff9bd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.238789 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e478a0b5-cea1-4639-a645-9660524ff9bd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.264308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7fr8\" (UniqueName: \"kubernetes.io/projected/e478a0b5-cea1-4639-a645-9660524ff9bd-kube-api-access-p7fr8\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.297206 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e478a0b5-cea1-4639-a645-9660524ff9bd\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.330200 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkmsb\" (UniqueName: \"kubernetes.io/projected/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-kube-api-access-xkmsb\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.330328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-config-data\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.330395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-combined-ca-bundle\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.330434 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-memcached-tls-certs\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.330489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-kolla-config\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.331442 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-kolla-config\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.332239 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-config-data\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.338683 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-memcached-tls-certs\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.339238 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-combined-ca-bundle\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.356909 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.357803 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkmsb\" (UniqueName: \"kubernetes.io/projected/91e88f07-d0d0-46d4-9e6a-a973b37b8a72-kube-api-access-xkmsb\") pod \"memcached-0\" (UID: \"91e88f07-d0d0-46d4-9e6a-a973b37b8a72\") " pod="openstack/memcached-0" Dec 04 14:16:42 crc kubenswrapper[4715]: I1204 14:16:42.456706 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 14:16:43 crc kubenswrapper[4715]: I1204 14:16:43.969292 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:16:43 crc kubenswrapper[4715]: I1204 14:16:43.970506 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:16:43 crc kubenswrapper[4715]: I1204 14:16:43.978531 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9857s" Dec 04 14:16:43 crc kubenswrapper[4715]: I1204 14:16:43.980860 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:16:44 crc kubenswrapper[4715]: I1204 14:16:44.058665 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkh7d\" (UniqueName: \"kubernetes.io/projected/435f8c50-e8e5-496f-9602-244bc202517f-kube-api-access-gkh7d\") pod \"kube-state-metrics-0\" (UID: \"435f8c50-e8e5-496f-9602-244bc202517f\") " pod="openstack/kube-state-metrics-0" Dec 04 14:16:44 crc kubenswrapper[4715]: I1204 14:16:44.160099 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkh7d\" (UniqueName: \"kubernetes.io/projected/435f8c50-e8e5-496f-9602-244bc202517f-kube-api-access-gkh7d\") pod \"kube-state-metrics-0\" (UID: \"435f8c50-e8e5-496f-9602-244bc202517f\") " pod="openstack/kube-state-metrics-0" Dec 04 14:16:44 crc kubenswrapper[4715]: I1204 14:16:44.180122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkh7d\" (UniqueName: \"kubernetes.io/projected/435f8c50-e8e5-496f-9602-244bc202517f-kube-api-access-gkh7d\") pod \"kube-state-metrics-0\" (UID: \"435f8c50-e8e5-496f-9602-244bc202517f\") " pod="openstack/kube-state-metrics-0" Dec 04 14:16:44 crc kubenswrapper[4715]: I1204 14:16:44.291721 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.525810 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.530193 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.532961 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.533152 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.533488 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.533704 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k6qww" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.533826 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.545300 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.584505 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6k4jg"] Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.629782 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-run-ovn\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q2sd\" (UniqueName: \"kubernetes.io/projected/00743321-f084-4e95-b47d-1b21332be3d5-kube-api-access-2q2sd\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00743321-f084-4e95-b47d-1b21332be3d5-combined-ca-bundle\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637845 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-log-ovn\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00743321-f084-4e95-b47d-1b21332be3d5-scripts\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637903 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-run\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.637921 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/00743321-f084-4e95-b47d-1b21332be3d5-ovn-controller-tls-certs\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.643201 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-w5bbn" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.643384 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.644127 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.644672 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k4jg"] Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.706696 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-vmdp8"] Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.708685 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.746345 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vmdp8"] Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.765725 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-run-ovn\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.765815 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q2sd\" (UniqueName: \"kubernetes.io/projected/00743321-f084-4e95-b47d-1b21332be3d5-kube-api-access-2q2sd\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.765906 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d693b193-c70b-4654-828d-5cd4744bee62-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hngk\" (UniqueName: \"kubernetes.io/projected/d693b193-c70b-4654-828d-5cd4744bee62-kube-api-access-4hngk\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766303 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d693b193-c70b-4654-828d-5cd4744bee62-config\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00743321-f084-4e95-b47d-1b21332be3d5-combined-ca-bundle\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766432 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766505 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-log-ovn\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766573 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00743321-f084-4e95-b47d-1b21332be3d5-scripts\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766688 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766764 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-run\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.766792 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/00743321-f084-4e95-b47d-1b21332be3d5-ovn-controller-tls-certs\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.774396 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-run-ovn\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.774428 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d693b193-c70b-4654-828d-5cd4744bee62-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.775110 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-log-ovn\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.775380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/00743321-f084-4e95-b47d-1b21332be3d5-var-run\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.782098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00743321-f084-4e95-b47d-1b21332be3d5-scripts\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.793246 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/00743321-f084-4e95-b47d-1b21332be3d5-ovn-controller-tls-certs\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.793914 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00743321-f084-4e95-b47d-1b21332be3d5-combined-ca-bundle\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.797679 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q2sd\" (UniqueName: \"kubernetes.io/projected/00743321-f084-4e95-b47d-1b21332be3d5-kube-api-access-2q2sd\") pod \"ovn-controller-6k4jg\" (UID: \"00743321-f084-4e95-b47d-1b21332be3d5\") " pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876636 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kv2t\" (UniqueName: \"kubernetes.io/projected/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-kube-api-access-9kv2t\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876743 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-run\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d693b193-c70b-4654-828d-5cd4744bee62-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876812 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-scripts\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876880 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-log\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876926 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d693b193-c70b-4654-828d-5cd4744bee62-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hngk\" (UniqueName: \"kubernetes.io/projected/d693b193-c70b-4654-828d-5cd4744bee62-kube-api-access-4hngk\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-lib\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.876980 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d693b193-c70b-4654-828d-5cd4744bee62-config\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.877006 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-etc-ovs\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.877091 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.877132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.877678 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d693b193-c70b-4654-828d-5cd4744bee62-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.877955 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.879870 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d693b193-c70b-4654-828d-5cd4744bee62-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.883500 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d693b193-c70b-4654-828d-5cd4744bee62-config\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.883596 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.884552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.889447 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d693b193-c70b-4654-828d-5cd4744bee62-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.907873 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.914247 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hngk\" (UniqueName: \"kubernetes.io/projected/d693b193-c70b-4654-828d-5cd4744bee62-kube-api-access-4hngk\") pod \"ovsdbserver-nb-0\" (UID: \"d693b193-c70b-4654-828d-5cd4744bee62\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.970761 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k4jg" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978567 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-scripts\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978642 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-log\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978681 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-lib\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978734 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-etc-ovs\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978797 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kv2t\" (UniqueName: \"kubernetes.io/projected/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-kube-api-access-9kv2t\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978822 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-run\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.978988 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-run\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.979232 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-etc-ovs\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.979277 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-lib\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.979548 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-var-log\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.981708 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-scripts\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:48 crc kubenswrapper[4715]: I1204 14:16:48.999144 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kv2t\" (UniqueName: \"kubernetes.io/projected/2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf-kube-api-access-9kv2t\") pod \"ovn-controller-ovs-vmdp8\" (UID: \"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf\") " pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:49 crc kubenswrapper[4715]: I1204 14:16:49.039259 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:16:49 crc kubenswrapper[4715]: I1204 14:16:49.164419 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.848159 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.849850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.853593 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.853809 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.853965 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7msjn" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.854135 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.864474 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.948877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch9vh\" (UniqueName: \"kubernetes.io/projected/f01e0a29-0003-41df-a5c9-0ede7a249acc-kube-api-access-ch9vh\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f01e0a29-0003-41df-a5c9-0ede7a249acc-config\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949320 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f01e0a29-0003-41df-a5c9-0ede7a249acc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949596 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949785 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f01e0a29-0003-41df-a5c9-0ede7a249acc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:51 crc kubenswrapper[4715]: I1204 14:16:51.949888 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.052067 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f01e0a29-0003-41df-a5c9-0ede7a249acc-config\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.052134 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f01e0a29-0003-41df-a5c9-0ede7a249acc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.052156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.052554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.052725 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.053023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f01e0a29-0003-41df-a5c9-0ede7a249acc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.053081 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.053135 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.053365 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch9vh\" (UniqueName: \"kubernetes.io/projected/f01e0a29-0003-41df-a5c9-0ede7a249acc-kube-api-access-ch9vh\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.053593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f01e0a29-0003-41df-a5c9-0ede7a249acc-config\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.053667 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f01e0a29-0003-41df-a5c9-0ede7a249acc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.054436 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f01e0a29-0003-41df-a5c9-0ede7a249acc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.070599 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.070810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.071585 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01e0a29-0003-41df-a5c9-0ede7a249acc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.072672 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch9vh\" (UniqueName: \"kubernetes.io/projected/f01e0a29-0003-41df-a5c9-0ede7a249acc-kube-api-access-ch9vh\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.095366 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f01e0a29-0003-41df-a5c9-0ede7a249acc\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:16:52 crc kubenswrapper[4715]: I1204 14:16:52.199460 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 14:17:00 crc kubenswrapper[4715]: I1204 14:17:00.156113 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.894474 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.894650 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zrt59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(d34885f9-cf98-4b1a-bbf3-c7af62f5f273): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.896050 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.979481 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.979837 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbv2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-s9hhd_openstack(e4cf7297-95ff-4dc0-ac99-528ff7225438): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.981156 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" podUID="e4cf7297-95ff-4dc0-ac99-528ff7225438" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.987175 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.987423 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gc97,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-stnmb_openstack(03459255-c53c-4f8e-a1e3-b1a849a21f85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:17:00 crc kubenswrapper[4715]: E1204 14:17:00.988602 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" podUID="03459255-c53c-4f8e-a1e3-b1a849a21f85" Dec 04 14:17:01 crc kubenswrapper[4715]: W1204 14:17:01.019736 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod435f8c50_e8e5_496f_9602_244bc202517f.slice/crio-1128bc67c2442d053044f35109a2f5944f41ecb4b885ddeb750c00e53f0a3caf WatchSource:0}: Error finding container 1128bc67c2442d053044f35109a2f5944f41ecb4b885ddeb750c00e53f0a3caf: Status 404 returned error can't find the container with id 1128bc67c2442d053044f35109a2f5944f41ecb4b885ddeb750c00e53f0a3caf Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.540408 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.651389 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.726735 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k4jg"] Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.832980 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435f8c50-e8e5-496f-9602-244bc202517f","Type":"ContainerStarted","Data":"1128bc67c2442d053044f35109a2f5944f41ecb4b885ddeb750c00e53f0a3caf"} Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.841509 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.843440 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerID="580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748" exitCode=0 Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.843700 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" event={"ID":"6d6533ec-6214-4aa6-828c-707f1d9f8f11","Type":"ContainerDied","Data":"580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748"} Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.845615 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ddf5478-4875-4285-98af-0739c4636797" containerID="33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6" exitCode=0 Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.845667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" event={"ID":"4ddf5478-4875-4285-98af-0739c4636797","Type":"ContainerDied","Data":"33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6"} Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.849002 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 14:17:01 crc kubenswrapper[4715]: I1204 14:17:01.850325 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7d666280-e175-4f31-8e11-7af2d5e27e5a","Type":"ContainerStarted","Data":"56c9f4d315352186bd23433f45ee45de9caefda3c6bc3b75929db752e05bcd39"} Dec 04 14:17:02 crc kubenswrapper[4715]: I1204 14:17:02.056704 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 14:17:02 crc kubenswrapper[4715]: I1204 14:17:02.122701 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vmdp8"] Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.004136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f01e0a29-0003-41df-a5c9-0ede7a249acc","Type":"ContainerStarted","Data":"e5ac37eba67abe221ec0bed7259dff3e85de5a5e3c4313417e6f98b328b8469d"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.011914 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vmdp8" event={"ID":"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf","Type":"ContainerStarted","Data":"e175b9ca7e945f4e25383ba3078f3f55d820a71c0ffff7091297bf1059800bb0"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.013887 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e478a0b5-cea1-4639-a645-9660524ff9bd","Type":"ContainerStarted","Data":"289a19beaeaba9e28b5881a900ec7df1f8c3e0db202106a961c8dc34c11e1d52"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.016780 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" event={"ID":"03459255-c53c-4f8e-a1e3-b1a849a21f85","Type":"ContainerDied","Data":"e616980656c160cfa10e3f1397f90a89a37d9e1f9299d3c7816b3269f6761e4d"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.016828 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e616980656c160cfa10e3f1397f90a89a37d9e1f9299d3c7816b3269f6761e4d" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.019776 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"440f7f8d-a89d-4d6d-8280-c56fede1c91c","Type":"ContainerStarted","Data":"70e712e943ad0c774cffe640698edc91acb2e0c08db4632f0be92d036d4b9251"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.021822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d693b193-c70b-4654-828d-5cd4744bee62","Type":"ContainerStarted","Data":"435f5779ccc8cbec5ce90a2babdb9a79a4f2153c7575273216cec08ae651d40d"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.022322 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.024050 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"91e88f07-d0d0-46d4-9e6a-a973b37b8a72","Type":"ContainerStarted","Data":"8557bcf8e8e4892fcb000ebdb8622f7585d88f30560e1a6d6a23db4048644f1c"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.027689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k4jg" event={"ID":"00743321-f084-4e95-b47d-1b21332be3d5","Type":"ContainerStarted","Data":"18238d646ad0f7b0283be253d78e95e32c258def502dac9d9b93a90a6f7dbc66"} Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.177708 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03459255-c53c-4f8e-a1e3-b1a849a21f85-config\") pod \"03459255-c53c-4f8e-a1e3-b1a849a21f85\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.177897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gc97\" (UniqueName: \"kubernetes.io/projected/03459255-c53c-4f8e-a1e3-b1a849a21f85-kube-api-access-9gc97\") pod \"03459255-c53c-4f8e-a1e3-b1a849a21f85\" (UID: \"03459255-c53c-4f8e-a1e3-b1a849a21f85\") " Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.178314 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03459255-c53c-4f8e-a1e3-b1a849a21f85-config" (OuterVolumeSpecName: "config") pod "03459255-c53c-4f8e-a1e3-b1a849a21f85" (UID: "03459255-c53c-4f8e-a1e3-b1a849a21f85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.179848 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03459255-c53c-4f8e-a1e3-b1a849a21f85-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.202893 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03459255-c53c-4f8e-a1e3-b1a849a21f85-kube-api-access-9gc97" (OuterVolumeSpecName: "kube-api-access-9gc97") pod "03459255-c53c-4f8e-a1e3-b1a849a21f85" (UID: "03459255-c53c-4f8e-a1e3-b1a849a21f85"). InnerVolumeSpecName "kube-api-access-9gc97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.282239 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gc97\" (UniqueName: \"kubernetes.io/projected/03459255-c53c-4f8e-a1e3-b1a849a21f85-kube-api-access-9gc97\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.490278 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.589303 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbv2q\" (UniqueName: \"kubernetes.io/projected/e4cf7297-95ff-4dc0-ac99-528ff7225438-kube-api-access-xbv2q\") pod \"e4cf7297-95ff-4dc0-ac99-528ff7225438\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.589388 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-config\") pod \"e4cf7297-95ff-4dc0-ac99-528ff7225438\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.589433 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-dns-svc\") pod \"e4cf7297-95ff-4dc0-ac99-528ff7225438\" (UID: \"e4cf7297-95ff-4dc0-ac99-528ff7225438\") " Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.589820 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4cf7297-95ff-4dc0-ac99-528ff7225438" (UID: "e4cf7297-95ff-4dc0-ac99-528ff7225438"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.590296 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-config" (OuterVolumeSpecName: "config") pod "e4cf7297-95ff-4dc0-ac99-528ff7225438" (UID: "e4cf7297-95ff-4dc0-ac99-528ff7225438"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.608231 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4cf7297-95ff-4dc0-ac99-528ff7225438-kube-api-access-xbv2q" (OuterVolumeSpecName: "kube-api-access-xbv2q") pod "e4cf7297-95ff-4dc0-ac99-528ff7225438" (UID: "e4cf7297-95ff-4dc0-ac99-528ff7225438"). InnerVolumeSpecName "kube-api-access-xbv2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.691390 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbv2q\" (UniqueName: \"kubernetes.io/projected/e4cf7297-95ff-4dc0-ac99-528ff7225438-kube-api-access-xbv2q\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.691434 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:03 crc kubenswrapper[4715]: I1204 14:17:03.691504 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4cf7297-95ff-4dc0-ac99-528ff7225438-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.042749 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.042791 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-s9hhd" event={"ID":"e4cf7297-95ff-4dc0-ac99-528ff7225438","Type":"ContainerDied","Data":"21280a0ed0181326a2595304151ec3d3b10c6663953d9ce62383e1686e98a39c"} Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.042823 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-stnmb" Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.123480 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-stnmb"] Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.137987 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-stnmb"] Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.160077 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s9hhd"] Dec 04 14:17:04 crc kubenswrapper[4715]: I1204 14:17:04.168002 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s9hhd"] Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.054558 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" event={"ID":"4ddf5478-4875-4285-98af-0739c4636797","Type":"ContainerStarted","Data":"083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d"} Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.054942 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.056560 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" event={"ID":"6d6533ec-6214-4aa6-828c-707f1d9f8f11","Type":"ContainerStarted","Data":"3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f"} Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.056696 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.107697 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" podStartSLOduration=5.444088376 podStartE2EDuration="28.107678787s" podCreationTimestamp="2025-12-04 14:16:37 +0000 UTC" firstStartedPulling="2025-12-04 14:16:38.461654806 +0000 UTC m=+1155.530373051" lastFinishedPulling="2025-12-04 14:17:01.125245247 +0000 UTC m=+1178.193963462" observedRunningTime="2025-12-04 14:17:05.081705924 +0000 UTC m=+1182.150424179" watchObservedRunningTime="2025-12-04 14:17:05.107678787 +0000 UTC m=+1182.176397002" Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.109325 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" podStartSLOduration=5.813930401 podStartE2EDuration="28.10931569s" podCreationTimestamp="2025-12-04 14:16:37 +0000 UTC" firstStartedPulling="2025-12-04 14:16:38.834742499 +0000 UTC m=+1155.903460714" lastFinishedPulling="2025-12-04 14:17:01.130127788 +0000 UTC m=+1178.198846003" observedRunningTime="2025-12-04 14:17:05.102832327 +0000 UTC m=+1182.171550572" watchObservedRunningTime="2025-12-04 14:17:05.10931569 +0000 UTC m=+1182.178033905" Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.199694 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03459255-c53c-4f8e-a1e3-b1a849a21f85" path="/var/lib/kubelet/pods/03459255-c53c-4f8e-a1e3-b1a849a21f85/volumes" Dec 04 14:17:05 crc kubenswrapper[4715]: I1204 14:17:05.200258 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4cf7297-95ff-4dc0-ac99-528ff7225438" path="/var/lib/kubelet/pods/e4cf7297-95ff-4dc0-ac99-528ff7225438/volumes" Dec 04 14:17:06 crc kubenswrapper[4715]: I1204 14:17:06.065473 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34885f9-cf98-4b1a-bbf3-c7af62f5f273","Type":"ContainerStarted","Data":"41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e"} Dec 04 14:17:12 crc kubenswrapper[4715]: I1204 14:17:12.892294 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.216578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435f8c50-e8e5-496f-9602-244bc202517f","Type":"ContainerStarted","Data":"5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.216753 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.218109 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vmdp8" event={"ID":"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf","Type":"ContainerStarted","Data":"5c1b9ce4e06d22752c8e6022076169cc3d470c4efe65f89680a33fa8baa49a93"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.220831 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e478a0b5-cea1-4639-a645-9660524ff9bd","Type":"ContainerStarted","Data":"96e2ff3ca1e607ed1d474bb906c920f3992bc606f8bf7c097fc76576916c30d0"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.223781 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d693b193-c70b-4654-828d-5cd4744bee62","Type":"ContainerStarted","Data":"ce0354ad30e1a0132e3e435f006bb314035c62ead118fb72dc868a804a163a2f"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.225467 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"91e88f07-d0d0-46d4-9e6a-a973b37b8a72","Type":"ContainerStarted","Data":"d89eec0e7cf6bc1ede2fafc40eb6986bff7a83e37740ddc5ed9833fb886d7b8d"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.225598 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.227217 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k4jg" event={"ID":"00743321-f084-4e95-b47d-1b21332be3d5","Type":"ContainerStarted","Data":"c080cc8d4ea8196961ad01cccd3088841e4ec12603954185adec88e198b389eb"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.227636 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6k4jg" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.229718 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7d666280-e175-4f31-8e11-7af2d5e27e5a","Type":"ContainerStarted","Data":"dc99a778b92b6ebc05364fdbfda30bb9829cdc96c15efd163deef8ea19790942"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.231160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f01e0a29-0003-41df-a5c9-0ede7a249acc","Type":"ContainerStarted","Data":"c325ca79fc0e5d2b56f6ba54dcd459ffcdf255354ecaa6a51ed780199a32a6f4"} Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.239883 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.244241 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.294809905 podStartE2EDuration="30.244225525s" podCreationTimestamp="2025-12-04 14:16:43 +0000 UTC" firstStartedPulling="2025-12-04 14:17:01.024025897 +0000 UTC m=+1178.092744112" lastFinishedPulling="2025-12-04 14:17:11.973441517 +0000 UTC m=+1189.042159732" observedRunningTime="2025-12-04 14:17:13.238280636 +0000 UTC m=+1190.306998861" watchObservedRunningTime="2025-12-04 14:17:13.244225525 +0000 UTC m=+1190.312943740" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.318221 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6k4jg" podStartSLOduration=15.173645837 podStartE2EDuration="25.318198298s" podCreationTimestamp="2025-12-04 14:16:48 +0000 UTC" firstStartedPulling="2025-12-04 14:17:01.829436021 +0000 UTC m=+1178.898154236" lastFinishedPulling="2025-12-04 14:17:11.973988472 +0000 UTC m=+1189.042706697" observedRunningTime="2025-12-04 14:17:13.310800211 +0000 UTC m=+1190.379518426" watchObservedRunningTime="2025-12-04 14:17:13.318198298 +0000 UTC m=+1190.386916523" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.341314 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=21.510923595 podStartE2EDuration="31.341292394s" podCreationTimestamp="2025-12-04 14:16:42 +0000 UTC" firstStartedPulling="2025-12-04 14:17:02.132764283 +0000 UTC m=+1179.201482498" lastFinishedPulling="2025-12-04 14:17:11.963133082 +0000 UTC m=+1189.031851297" observedRunningTime="2025-12-04 14:17:13.332736366 +0000 UTC m=+1190.401454601" watchObservedRunningTime="2025-12-04 14:17:13.341292394 +0000 UTC m=+1190.410010609" Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.416095 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-66hsn"] Dec 04 14:17:13 crc kubenswrapper[4715]: I1204 14:17:13.416301 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" podUID="4ddf5478-4875-4285-98af-0739c4636797" containerName="dnsmasq-dns" containerID="cri-o://083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d" gracePeriod=10 Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.008589 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.068328 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-config\") pod \"4ddf5478-4875-4285-98af-0739c4636797\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.068383 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtmwd\" (UniqueName: \"kubernetes.io/projected/4ddf5478-4875-4285-98af-0739c4636797-kube-api-access-gtmwd\") pod \"4ddf5478-4875-4285-98af-0739c4636797\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.068490 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-dns-svc\") pod \"4ddf5478-4875-4285-98af-0739c4636797\" (UID: \"4ddf5478-4875-4285-98af-0739c4636797\") " Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.073790 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddf5478-4875-4285-98af-0739c4636797-kube-api-access-gtmwd" (OuterVolumeSpecName: "kube-api-access-gtmwd") pod "4ddf5478-4875-4285-98af-0739c4636797" (UID: "4ddf5478-4875-4285-98af-0739c4636797"). InnerVolumeSpecName "kube-api-access-gtmwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.106864 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ddf5478-4875-4285-98af-0739c4636797" (UID: "4ddf5478-4875-4285-98af-0739c4636797"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.119797 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-config" (OuterVolumeSpecName: "config") pod "4ddf5478-4875-4285-98af-0739c4636797" (UID: "4ddf5478-4875-4285-98af-0739c4636797"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.171132 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.171167 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtmwd\" (UniqueName: \"kubernetes.io/projected/4ddf5478-4875-4285-98af-0739c4636797-kube-api-access-gtmwd\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.171178 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddf5478-4875-4285-98af-0739c4636797-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.242319 4715 generic.go:334] "Generic (PLEG): container finished" podID="2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf" containerID="5c1b9ce4e06d22752c8e6022076169cc3d470c4efe65f89680a33fa8baa49a93" exitCode=0 Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.242398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vmdp8" event={"ID":"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf","Type":"ContainerDied","Data":"5c1b9ce4e06d22752c8e6022076169cc3d470c4efe65f89680a33fa8baa49a93"} Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.245242 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ddf5478-4875-4285-98af-0739c4636797" containerID="083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d" exitCode=0 Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.246223 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.246267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" event={"ID":"4ddf5478-4875-4285-98af-0739c4636797","Type":"ContainerDied","Data":"083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d"} Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.246300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-66hsn" event={"ID":"4ddf5478-4875-4285-98af-0739c4636797","Type":"ContainerDied","Data":"ccfaaf41cccdd5e5ac1e94651b2d0c359967d4c3c033be38f26c85345fb1be05"} Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.246321 4715 scope.go:117] "RemoveContainer" containerID="083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.281902 4715 scope.go:117] "RemoveContainer" containerID="33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.285858 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-66hsn"] Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.292367 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-66hsn"] Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.323303 4715 scope.go:117] "RemoveContainer" containerID="083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d" Dec 04 14:17:14 crc kubenswrapper[4715]: E1204 14:17:14.323769 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d\": container with ID starting with 083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d not found: ID does not exist" containerID="083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.323804 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d"} err="failed to get container status \"083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d\": rpc error: code = NotFound desc = could not find container \"083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d\": container with ID starting with 083c1e6e37615416050f8fcb20bfaaaa89b94ee957961a934ed779434860b53d not found: ID does not exist" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.323832 4715 scope.go:117] "RemoveContainer" containerID="33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6" Dec 04 14:17:14 crc kubenswrapper[4715]: E1204 14:17:14.324323 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6\": container with ID starting with 33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6 not found: ID does not exist" containerID="33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6" Dec 04 14:17:14 crc kubenswrapper[4715]: I1204 14:17:14.324349 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6"} err="failed to get container status \"33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6\": rpc error: code = NotFound desc = could not find container \"33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6\": container with ID starting with 33806acfdab3e2864f7da69d1214abd6d8b9875346ab76376de8f34e9d8320d6 not found: ID does not exist" Dec 04 14:17:15 crc kubenswrapper[4715]: I1204 14:17:15.193977 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ddf5478-4875-4285-98af-0739c4636797" path="/var/lib/kubelet/pods/4ddf5478-4875-4285-98af-0739c4636797/volumes" Dec 04 14:17:15 crc kubenswrapper[4715]: I1204 14:17:15.259517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vmdp8" event={"ID":"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf","Type":"ContainerStarted","Data":"d035c4e4e66358ab3bafaa79c23970259d5b5c5535f3bc45b73b81e5eb213a9c"} Dec 04 14:17:15 crc kubenswrapper[4715]: I1204 14:17:15.260666 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vmdp8" event={"ID":"2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf","Type":"ContainerStarted","Data":"58b2a0f0ce22dfa3bdef0c41e344f86f895e7039f78a018a10c337ed185f9a93"} Dec 04 14:17:15 crc kubenswrapper[4715]: I1204 14:17:15.260945 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:17:15 crc kubenswrapper[4715]: I1204 14:17:15.291738 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-vmdp8" podStartSLOduration=17.509483504 podStartE2EDuration="27.29171626s" podCreationTimestamp="2025-12-04 14:16:48 +0000 UTC" firstStartedPulling="2025-12-04 14:17:02.191940051 +0000 UTC m=+1179.260658256" lastFinishedPulling="2025-12-04 14:17:11.974172797 +0000 UTC m=+1189.042891012" observedRunningTime="2025-12-04 14:17:15.280114551 +0000 UTC m=+1192.348832786" watchObservedRunningTime="2025-12-04 14:17:15.29171626 +0000 UTC m=+1192.360434475" Dec 04 14:17:16 crc kubenswrapper[4715]: I1204 14:17:16.285979 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.294945 4715 generic.go:334] "Generic (PLEG): container finished" podID="e478a0b5-cea1-4639-a645-9660524ff9bd" containerID="96e2ff3ca1e607ed1d474bb906c920f3992bc606f8bf7c097fc76576916c30d0" exitCode=0 Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.295058 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e478a0b5-cea1-4639-a645-9660524ff9bd","Type":"ContainerDied","Data":"96e2ff3ca1e607ed1d474bb906c920f3992bc606f8bf7c097fc76576916c30d0"} Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.298818 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d693b193-c70b-4654-828d-5cd4744bee62","Type":"ContainerStarted","Data":"8684b08d37d221876f341cc14fb8a5dd000a9dd9ad85ed167c29a5ccd7719878"} Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.300868 4715 generic.go:334] "Generic (PLEG): container finished" podID="7d666280-e175-4f31-8e11-7af2d5e27e5a" containerID="dc99a778b92b6ebc05364fdbfda30bb9829cdc96c15efd163deef8ea19790942" exitCode=0 Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.300948 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7d666280-e175-4f31-8e11-7af2d5e27e5a","Type":"ContainerDied","Data":"dc99a778b92b6ebc05364fdbfda30bb9829cdc96c15efd163deef8ea19790942"} Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.308542 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f01e0a29-0003-41df-a5c9-0ede7a249acc","Type":"ContainerStarted","Data":"3815a691c69a8193446ce6c50ce397e8e560af6272fed3339790adbad293af06"} Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.384811 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=15.874631111 podStartE2EDuration="30.384788052s" podCreationTimestamp="2025-12-04 14:16:47 +0000 UTC" firstStartedPulling="2025-12-04 14:17:01.829160864 +0000 UTC m=+1178.897879079" lastFinishedPulling="2025-12-04 14:17:16.339317805 +0000 UTC m=+1193.408036020" observedRunningTime="2025-12-04 14:17:17.348317639 +0000 UTC m=+1194.417035874" watchObservedRunningTime="2025-12-04 14:17:17.384788052 +0000 UTC m=+1194.453506257" Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.405806 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.271457507 podStartE2EDuration="27.405784672s" podCreationTimestamp="2025-12-04 14:16:50 +0000 UTC" firstStartedPulling="2025-12-04 14:17:02.19338857 +0000 UTC m=+1179.262106785" lastFinishedPulling="2025-12-04 14:17:16.327715725 +0000 UTC m=+1193.396433950" observedRunningTime="2025-12-04 14:17:17.399782122 +0000 UTC m=+1194.468500337" watchObservedRunningTime="2025-12-04 14:17:17.405784672 +0000 UTC m=+1194.474502887" Dec 04 14:17:17 crc kubenswrapper[4715]: I1204 14:17:17.507178 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 14:17:18 crc kubenswrapper[4715]: I1204 14:17:18.317352 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7d666280-e175-4f31-8e11-7af2d5e27e5a","Type":"ContainerStarted","Data":"fcf0838c0bd23ea830adbadb9c5232395ce167e22c323290b40ec20352582712"} Dec 04 14:17:18 crc kubenswrapper[4715]: I1204 14:17:18.319548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e478a0b5-cea1-4639-a645-9660524ff9bd","Type":"ContainerStarted","Data":"44838110c40ad6af62109d00eb14a65fe4f2678c13d931c4292ccd936ced4c29"} Dec 04 14:17:18 crc kubenswrapper[4715]: I1204 14:17:18.343222 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=28.93145195 podStartE2EDuration="39.343201347s" podCreationTimestamp="2025-12-04 14:16:39 +0000 UTC" firstStartedPulling="2025-12-04 14:17:01.55158272 +0000 UTC m=+1178.620300935" lastFinishedPulling="2025-12-04 14:17:11.963332117 +0000 UTC m=+1189.032050332" observedRunningTime="2025-12-04 14:17:18.337425693 +0000 UTC m=+1195.406143908" watchObservedRunningTime="2025-12-04 14:17:18.343201347 +0000 UTC m=+1195.411919572" Dec 04 14:17:18 crc kubenswrapper[4715]: I1204 14:17:18.358545 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.589601586 podStartE2EDuration="37.358523816s" podCreationTimestamp="2025-12-04 14:16:41 +0000 UTC" firstStartedPulling="2025-12-04 14:17:02.192090555 +0000 UTC m=+1179.260808780" lastFinishedPulling="2025-12-04 14:17:11.961012795 +0000 UTC m=+1189.029731010" observedRunningTime="2025-12-04 14:17:18.354591671 +0000 UTC m=+1195.423309896" watchObservedRunningTime="2025-12-04 14:17:18.358523816 +0000 UTC m=+1195.427242051" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.165966 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.166013 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.200180 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.210390 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.248523 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.328425 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.365723 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.372182 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.555434 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-258ws"] Dec 04 14:17:19 crc kubenswrapper[4715]: E1204 14:17:19.555916 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddf5478-4875-4285-98af-0739c4636797" containerName="dnsmasq-dns" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.555935 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddf5478-4875-4285-98af-0739c4636797" containerName="dnsmasq-dns" Dec 04 14:17:19 crc kubenswrapper[4715]: E1204 14:17:19.555943 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddf5478-4875-4285-98af-0739c4636797" containerName="init" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.555949 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddf5478-4875-4285-98af-0739c4636797" containerName="init" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.556144 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddf5478-4875-4285-98af-0739c4636797" containerName="dnsmasq-dns" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.557233 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.559406 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.584994 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-258ws"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.674603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fs74\" (UniqueName: \"kubernetes.io/projected/462af9dd-9e6d-4614-9633-a1fbd558c9e5-kube-api-access-4fs74\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.674705 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.674776 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-config\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.674815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.722995 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-tw9p4"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.724093 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.729222 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.748679 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-tw9p4"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.777159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.777244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-config\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.777277 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.777328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fs74\" (UniqueName: \"kubernetes.io/projected/462af9dd-9e6d-4614-9633-a1fbd558c9e5-kube-api-access-4fs74\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.778175 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.778215 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-config\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.778449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.823334 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fs74\" (UniqueName: \"kubernetes.io/projected/462af9dd-9e6d-4614-9633-a1fbd558c9e5-kube-api-access-4fs74\") pod \"dnsmasq-dns-7f896c8c65-258ws\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.841109 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.842541 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.846414 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-4657k" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.846674 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.846748 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.848872 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.855160 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-258ws"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.857725 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.866617 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.883858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44a78956-6eb6-43e2-98f1-2e2c15342fd8-combined-ca-bundle\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.883913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2h77\" (UniqueName: \"kubernetes.io/projected/44a78956-6eb6-43e2-98f1-2e2c15342fd8-kube-api-access-s2h77\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.883935 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a78956-6eb6-43e2-98f1-2e2c15342fd8-config\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.883959 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/44a78956-6eb6-43e2-98f1-2e2c15342fd8-ovs-rundir\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.884057 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/44a78956-6eb6-43e2-98f1-2e2c15342fd8-ovn-rundir\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.884079 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/44a78956-6eb6-43e2-98f1-2e2c15342fd8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.897839 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-79kj9"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.899501 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.903476 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.936562 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-79kj9"] Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-config\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985181 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc7b2\" (UniqueName: \"kubernetes.io/projected/f92de975-e892-45d7-86b9-45dd9e4115b5-kube-api-access-lc7b2\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985229 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/44a78956-6eb6-43e2-98f1-2e2c15342fd8-ovn-rundir\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985301 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57e4b94e-20df-457c-81e8-9c114b18bbe6-scripts\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985330 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/44a78956-6eb6-43e2-98f1-2e2c15342fd8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985360 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985407 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44a78956-6eb6-43e2-98f1-2e2c15342fd8-combined-ca-bundle\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985453 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e4b94e-20df-457c-81e8-9c114b18bbe6-config\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985479 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2h77\" (UniqueName: \"kubernetes.io/projected/44a78956-6eb6-43e2-98f1-2e2c15342fd8-kube-api-access-s2h77\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a78956-6eb6-43e2-98f1-2e2c15342fd8-config\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxqpq\" (UniqueName: \"kubernetes.io/projected/57e4b94e-20df-457c-81e8-9c114b18bbe6-kube-api-access-lxqpq\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985605 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57e4b94e-20df-457c-81e8-9c114b18bbe6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985626 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/44a78956-6eb6-43e2-98f1-2e2c15342fd8-ovs-rundir\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.985646 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.986411 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/44a78956-6eb6-43e2-98f1-2e2c15342fd8-ovn-rundir\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.989489 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/44a78956-6eb6-43e2-98f1-2e2c15342fd8-ovs-rundir\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.989546 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a78956-6eb6-43e2-98f1-2e2c15342fd8-config\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.993270 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/44a78956-6eb6-43e2-98f1-2e2c15342fd8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:19 crc kubenswrapper[4715]: I1204 14:17:19.993830 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44a78956-6eb6-43e2-98f1-2e2c15342fd8-combined-ca-bundle\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.021810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2h77\" (UniqueName: \"kubernetes.io/projected/44a78956-6eb6-43e2-98f1-2e2c15342fd8-kube-api-access-s2h77\") pod \"ovn-controller-metrics-tw9p4\" (UID: \"44a78956-6eb6-43e2-98f1-2e2c15342fd8\") " pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.041134 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tw9p4" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087518 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57e4b94e-20df-457c-81e8-9c114b18bbe6-scripts\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087630 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087702 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e4b94e-20df-457c-81e8-9c114b18bbe6-config\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087750 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087773 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxqpq\" (UniqueName: \"kubernetes.io/projected/57e4b94e-20df-457c-81e8-9c114b18bbe6-kube-api-access-lxqpq\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57e4b94e-20df-457c-81e8-9c114b18bbe6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087823 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087893 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-config\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.087924 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc7b2\" (UniqueName: \"kubernetes.io/projected/f92de975-e892-45d7-86b9-45dd9e4115b5-kube-api-access-lc7b2\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.088800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57e4b94e-20df-457c-81e8-9c114b18bbe6-scripts\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.089549 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.089702 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e4b94e-20df-457c-81e8-9c114b18bbe6-config\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.090168 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57e4b94e-20df-457c-81e8-9c114b18bbe6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.091026 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.091070 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.091609 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-config\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.094475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.096070 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.106694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e4b94e-20df-457c-81e8-9c114b18bbe6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.107339 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc7b2\" (UniqueName: \"kubernetes.io/projected/f92de975-e892-45d7-86b9-45dd9e4115b5-kube-api-access-lc7b2\") pod \"dnsmasq-dns-86db49b7ff-79kj9\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.109066 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxqpq\" (UniqueName: \"kubernetes.io/projected/57e4b94e-20df-457c-81e8-9c114b18bbe6-kube-api-access-lxqpq\") pod \"ovn-northd-0\" (UID: \"57e4b94e-20df-457c-81e8-9c114b18bbe6\") " pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.173740 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.219026 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.645996 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-258ws"] Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.692930 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-tw9p4"] Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.931267 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.931562 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 14:17:20 crc kubenswrapper[4715]: I1204 14:17:20.987056 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.088423 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-79kj9"] Dec 04 14:17:21 crc kubenswrapper[4715]: E1204 14:17:21.114193 4715 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.198:55644->38.102.83.198:41743: write tcp 38.102.83.198:55644->38.102.83.198:41743: write: broken pipe Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.606667 4715 generic.go:334] "Generic (PLEG): container finished" podID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerID="a350c8b672df37f357e4f0ad7859a13a7ec729f5524dc3e5406cf053fd77508c" exitCode=0 Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.606780 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" event={"ID":"f92de975-e892-45d7-86b9-45dd9e4115b5","Type":"ContainerDied","Data":"a350c8b672df37f357e4f0ad7859a13a7ec729f5524dc3e5406cf053fd77508c"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.607136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" event={"ID":"f92de975-e892-45d7-86b9-45dd9e4115b5","Type":"ContainerStarted","Data":"378183a29fbc6bdb42b414ea59ee5c2092905aadfcaa2dfa600c45aaa901883b"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.611212 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tw9p4" event={"ID":"44a78956-6eb6-43e2-98f1-2e2c15342fd8","Type":"ContainerStarted","Data":"ac8dbef06d88f5d14a84b75e2964d21e73ad56369fb9cee301fb6a3dfd8f4288"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.611252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tw9p4" event={"ID":"44a78956-6eb6-43e2-98f1-2e2c15342fd8","Type":"ContainerStarted","Data":"d25412eaacbd6938ec3cb9cfc13fde2746365cf7ae1e8d048977c2d84f7d8adb"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.612826 4715 generic.go:334] "Generic (PLEG): container finished" podID="462af9dd-9e6d-4614-9633-a1fbd558c9e5" containerID="eb054c9911bd552392ab2d9b3021a2c8eecb889184d4f0843f6882d17288725d" exitCode=0 Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.612881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" event={"ID":"462af9dd-9e6d-4614-9633-a1fbd558c9e5","Type":"ContainerDied","Data":"eb054c9911bd552392ab2d9b3021a2c8eecb889184d4f0843f6882d17288725d"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.612903 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" event={"ID":"462af9dd-9e6d-4614-9633-a1fbd558c9e5","Type":"ContainerStarted","Data":"66cf8e828ad02323138de117e825905d5adc6823f5231d509d1a56d7f06635f3"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.615374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"57e4b94e-20df-457c-81e8-9c114b18bbe6","Type":"ContainerStarted","Data":"168de89e7e12f05e356553cbf72eea3d8701c462852cc20975065924878154a4"} Dec 04 14:17:21 crc kubenswrapper[4715]: I1204 14:17:21.700580 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-tw9p4" podStartSLOduration=2.700557673 podStartE2EDuration="2.700557673s" podCreationTimestamp="2025-12-04 14:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:21.688199643 +0000 UTC m=+1198.756917858" watchObservedRunningTime="2025-12-04 14:17:21.700557673 +0000 UTC m=+1198.769275888" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.090184 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.133545 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-ovsdbserver-sb\") pod \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.133825 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fs74\" (UniqueName: \"kubernetes.io/projected/462af9dd-9e6d-4614-9633-a1fbd558c9e5-kube-api-access-4fs74\") pod \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.134130 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-dns-svc\") pod \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.134337 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-config\") pod \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\" (UID: \"462af9dd-9e6d-4614-9633-a1fbd558c9e5\") " Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.137669 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462af9dd-9e6d-4614-9633-a1fbd558c9e5-kube-api-access-4fs74" (OuterVolumeSpecName: "kube-api-access-4fs74") pod "462af9dd-9e6d-4614-9633-a1fbd558c9e5" (UID: "462af9dd-9e6d-4614-9633-a1fbd558c9e5"). InnerVolumeSpecName "kube-api-access-4fs74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.158370 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "462af9dd-9e6d-4614-9633-a1fbd558c9e5" (UID: "462af9dd-9e6d-4614-9633-a1fbd558c9e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.160095 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "462af9dd-9e6d-4614-9633-a1fbd558c9e5" (UID: "462af9dd-9e6d-4614-9633-a1fbd558c9e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.164017 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-config" (OuterVolumeSpecName: "config") pod "462af9dd-9e6d-4614-9633-a1fbd558c9e5" (UID: "462af9dd-9e6d-4614-9633-a1fbd558c9e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.236523 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.236565 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fs74\" (UniqueName: \"kubernetes.io/projected/462af9dd-9e6d-4614-9633-a1fbd558c9e5-kube-api-access-4fs74\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.236578 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.236589 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462af9dd-9e6d-4614-9633-a1fbd558c9e5-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.357815 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.358110 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.631657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" event={"ID":"f92de975-e892-45d7-86b9-45dd9e4115b5","Type":"ContainerStarted","Data":"d12875433248aea72fba337a5e13556dd5b49728876f9762caae11b9604f521d"} Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.631827 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.639564 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" event={"ID":"462af9dd-9e6d-4614-9633-a1fbd558c9e5","Type":"ContainerDied","Data":"66cf8e828ad02323138de117e825905d5adc6823f5231d509d1a56d7f06635f3"} Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.639615 4715 scope.go:117] "RemoveContainer" containerID="eb054c9911bd552392ab2d9b3021a2c8eecb889184d4f0843f6882d17288725d" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.639624 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-258ws" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.651864 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"57e4b94e-20df-457c-81e8-9c114b18bbe6","Type":"ContainerStarted","Data":"d71072b8e98dda4c1adff3b305762607883a86e8f2e118f0915f82b85c88ea4b"} Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.651914 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"57e4b94e-20df-457c-81e8-9c114b18bbe6","Type":"ContainerStarted","Data":"2b6cb515975cc9fe4abc14c95d95abd45d0590487f51c4268c3328c18a705060"} Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.652095 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.657519 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" podStartSLOduration=3.657493719 podStartE2EDuration="3.657493719s" podCreationTimestamp="2025-12-04 14:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:22.649785604 +0000 UTC m=+1199.718503819" watchObservedRunningTime="2025-12-04 14:17:22.657493719 +0000 UTC m=+1199.726211954" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.676108 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.598372577 podStartE2EDuration="3.676091235s" podCreationTimestamp="2025-12-04 14:17:19 +0000 UTC" firstStartedPulling="2025-12-04 14:17:21.012217922 +0000 UTC m=+1198.080936137" lastFinishedPulling="2025-12-04 14:17:22.08993658 +0000 UTC m=+1199.158654795" observedRunningTime="2025-12-04 14:17:22.674224116 +0000 UTC m=+1199.742942331" watchObservedRunningTime="2025-12-04 14:17:22.676091235 +0000 UTC m=+1199.744809450" Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.704138 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-258ws"] Dec 04 14:17:22 crc kubenswrapper[4715]: I1204 14:17:22.721766 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-258ws"] Dec 04 14:17:23 crc kubenswrapper[4715]: I1204 14:17:23.210729 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462af9dd-9e6d-4614-9633-a1fbd558c9e5" path="/var/lib/kubelet/pods/462af9dd-9e6d-4614-9633-a1fbd558c9e5/volumes" Dec 04 14:17:23 crc kubenswrapper[4715]: I1204 14:17:23.248363 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 14:17:23 crc kubenswrapper[4715]: I1204 14:17:23.331617 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.224559 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-79kj9"] Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.257984 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-tmg6v"] Dec 04 14:17:24 crc kubenswrapper[4715]: E1204 14:17:24.258502 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462af9dd-9e6d-4614-9633-a1fbd558c9e5" containerName="init" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.258528 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="462af9dd-9e6d-4614-9633-a1fbd558c9e5" containerName="init" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.258768 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="462af9dd-9e6d-4614-9633-a1fbd558c9e5" containerName="init" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.259903 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.278634 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tmg6v"] Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.314979 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.393329 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.393719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fxzp\" (UniqueName: \"kubernetes.io/projected/332031a0-392b-4497-9c04-9c1bbd0a2ea0-kube-api-access-6fxzp\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.393876 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-dns-svc\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.394049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.394250 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.495641 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.495696 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.495737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fxzp\" (UniqueName: \"kubernetes.io/projected/332031a0-392b-4497-9c04-9c1bbd0a2ea0-kube-api-access-6fxzp\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.495795 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-dns-svc\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.495835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.496943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.497056 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.496957 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.497186 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-dns-svc\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.520316 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fxzp\" (UniqueName: \"kubernetes.io/projected/332031a0-392b-4497-9c04-9c1bbd0a2ea0-kube-api-access-6fxzp\") pod \"dnsmasq-dns-698758b865-tmg6v\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.578224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:24 crc kubenswrapper[4715]: I1204 14:17:24.666149 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerName="dnsmasq-dns" containerID="cri-o://d12875433248aea72fba337a5e13556dd5b49728876f9762caae11b9604f521d" gracePeriod=10 Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.060598 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tmg6v"] Dec 04 14:17:25 crc kubenswrapper[4715]: W1204 14:17:25.133240 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod332031a0_392b_4497_9c04_9c1bbd0a2ea0.slice/crio-2487bb0491b566f50604f4d77385a0b8ebb9a717da3a7a79331955e87581c950 WatchSource:0}: Error finding container 2487bb0491b566f50604f4d77385a0b8ebb9a717da3a7a79331955e87581c950: Status 404 returned error can't find the container with id 2487bb0491b566f50604f4d77385a0b8ebb9a717da3a7a79331955e87581c950 Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.276700 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.378198 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.401521 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.408233 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.410275 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.411430 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.411458 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.412821 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-kkk4c" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.440664 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.516283 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/548d91b6-2482-4624-843a-ea2898f7beb8-lock\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.516530 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.516583 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/548d91b6-2482-4624-843a-ea2898f7beb8-cache\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.524027 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b9jr\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-kube-api-access-6b9jr\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.524224 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.626364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/548d91b6-2482-4624-843a-ea2898f7beb8-lock\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.626673 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.626718 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/548d91b6-2482-4624-843a-ea2898f7beb8-cache\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.626753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b9jr\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-kube-api-access-6b9jr\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.626869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/548d91b6-2482-4624-843a-ea2898f7beb8-lock\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: E1204 14:17:25.627005 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:17:25 crc kubenswrapper[4715]: E1204 14:17:25.627179 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:17:25 crc kubenswrapper[4715]: E1204 14:17:25.627239 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift podName:548d91b6-2482-4624-843a-ea2898f7beb8 nodeName:}" failed. No retries permitted until 2025-12-04 14:17:26.127223926 +0000 UTC m=+1203.195942141 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift") pod "swift-storage-0" (UID: "548d91b6-2482-4624-843a-ea2898f7beb8") : configmap "swift-ring-files" not found Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.627071 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/548d91b6-2482-4624-843a-ea2898f7beb8-cache\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.627389 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.627548 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.648847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b9jr\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-kube-api-access-6b9jr\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.651301 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.677747 4715 generic.go:334] "Generic (PLEG): container finished" podID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerID="d12875433248aea72fba337a5e13556dd5b49728876f9762caae11b9604f521d" exitCode=0 Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.677762 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" event={"ID":"f92de975-e892-45d7-86b9-45dd9e4115b5","Type":"ContainerDied","Data":"d12875433248aea72fba337a5e13556dd5b49728876f9762caae11b9604f521d"} Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.678115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" event={"ID":"f92de975-e892-45d7-86b9-45dd9e4115b5","Type":"ContainerDied","Data":"378183a29fbc6bdb42b414ea59ee5c2092905aadfcaa2dfa600c45aaa901883b"} Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.678190 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="378183a29fbc6bdb42b414ea59ee5c2092905aadfcaa2dfa600c45aaa901883b" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.680775 4715 generic.go:334] "Generic (PLEG): container finished" podID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerID="e47a95d9a58803cf79875408ee672b661f48a6df5064e2c0c6065d030cfcfbda" exitCode=0 Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.680892 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tmg6v" event={"ID":"332031a0-392b-4497-9c04-9c1bbd0a2ea0","Type":"ContainerDied","Data":"e47a95d9a58803cf79875408ee672b661f48a6df5064e2c0c6065d030cfcfbda"} Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.680976 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tmg6v" event={"ID":"332031a0-392b-4497-9c04-9c1bbd0a2ea0","Type":"ContainerStarted","Data":"2487bb0491b566f50604f4d77385a0b8ebb9a717da3a7a79331955e87581c950"} Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.828546 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.933869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-dns-svc\") pod \"f92de975-e892-45d7-86b9-45dd9e4115b5\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.934018 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-config\") pod \"f92de975-e892-45d7-86b9-45dd9e4115b5\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.934115 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-sb\") pod \"f92de975-e892-45d7-86b9-45dd9e4115b5\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.934183 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc7b2\" (UniqueName: \"kubernetes.io/projected/f92de975-e892-45d7-86b9-45dd9e4115b5-kube-api-access-lc7b2\") pod \"f92de975-e892-45d7-86b9-45dd9e4115b5\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.934260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-nb\") pod \"f92de975-e892-45d7-86b9-45dd9e4115b5\" (UID: \"f92de975-e892-45d7-86b9-45dd9e4115b5\") " Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.937580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92de975-e892-45d7-86b9-45dd9e4115b5-kube-api-access-lc7b2" (OuterVolumeSpecName: "kube-api-access-lc7b2") pod "f92de975-e892-45d7-86b9-45dd9e4115b5" (UID: "f92de975-e892-45d7-86b9-45dd9e4115b5"). InnerVolumeSpecName "kube-api-access-lc7b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.969255 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f92de975-e892-45d7-86b9-45dd9e4115b5" (UID: "f92de975-e892-45d7-86b9-45dd9e4115b5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.971098 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f92de975-e892-45d7-86b9-45dd9e4115b5" (UID: "f92de975-e892-45d7-86b9-45dd9e4115b5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.971715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-config" (OuterVolumeSpecName: "config") pod "f92de975-e892-45d7-86b9-45dd9e4115b5" (UID: "f92de975-e892-45d7-86b9-45dd9e4115b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:25 crc kubenswrapper[4715]: I1204 14:17:25.972589 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f92de975-e892-45d7-86b9-45dd9e4115b5" (UID: "f92de975-e892-45d7-86b9-45dd9e4115b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.036155 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.036197 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.036209 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.036221 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc7b2\" (UniqueName: \"kubernetes.io/projected/f92de975-e892-45d7-86b9-45dd9e4115b5-kube-api-access-lc7b2\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.036230 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f92de975-e892-45d7-86b9-45dd9e4115b5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.039406 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-5r72w"] Dec 04 14:17:26 crc kubenswrapper[4715]: E1204 14:17:26.039762 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerName="init" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.039777 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerName="init" Dec 04 14:17:26 crc kubenswrapper[4715]: E1204 14:17:26.039792 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerName="dnsmasq-dns" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.039798 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerName="dnsmasq-dns" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.039973 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" containerName="dnsmasq-dns" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.040558 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.043190 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.043289 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.043401 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.059666 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5r72w"] Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.137770 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d46h\" (UniqueName: \"kubernetes.io/projected/2386cbfc-2eba-4ace-a56a-0e203b7b9229-kube-api-access-6d46h\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.137819 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-combined-ca-bundle\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.137839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2386cbfc-2eba-4ace-a56a-0e203b7b9229-etc-swift\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.137872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-swiftconf\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.138020 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-ring-data-devices\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.138299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.138370 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-scripts\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.138447 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-dispersionconf\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: E1204 14:17:26.138656 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:17:26 crc kubenswrapper[4715]: E1204 14:17:26.138736 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:17:26 crc kubenswrapper[4715]: E1204 14:17:26.138842 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift podName:548d91b6-2482-4624-843a-ea2898f7beb8 nodeName:}" failed. No retries permitted until 2025-12-04 14:17:27.138810762 +0000 UTC m=+1204.207528977 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift") pod "swift-storage-0" (UID: "548d91b6-2482-4624-843a-ea2898f7beb8") : configmap "swift-ring-files" not found Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240272 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-dispersionconf\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240621 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d46h\" (UniqueName: \"kubernetes.io/projected/2386cbfc-2eba-4ace-a56a-0e203b7b9229-kube-api-access-6d46h\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-combined-ca-bundle\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240684 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2386cbfc-2eba-4ace-a56a-0e203b7b9229-etc-swift\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-swiftconf\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240765 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-ring-data-devices\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.240885 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-scripts\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.241835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-scripts\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.241867 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-ring-data-devices\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.242105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2386cbfc-2eba-4ace-a56a-0e203b7b9229-etc-swift\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.244066 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-swiftconf\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.244464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-dispersionconf\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.244861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-combined-ca-bundle\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.258380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d46h\" (UniqueName: \"kubernetes.io/projected/2386cbfc-2eba-4ace-a56a-0e203b7b9229-kube-api-access-6d46h\") pod \"swift-ring-rebalance-5r72w\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.378742 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.691465 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-79kj9" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.693318 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tmg6v" event={"ID":"332031a0-392b-4497-9c04-9c1bbd0a2ea0","Type":"ContainerStarted","Data":"855aabe107d7d0ce74853d7a1bb68811a7a508c6ed41cc67583881d3d58fcfe9"} Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.693368 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.713103 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-tmg6v" podStartSLOduration=2.7130794209999998 podStartE2EDuration="2.713079421s" podCreationTimestamp="2025-12-04 14:17:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:26.711394026 +0000 UTC m=+1203.780112261" watchObservedRunningTime="2025-12-04 14:17:26.713079421 +0000 UTC m=+1203.781797636" Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.734390 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-79kj9"] Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.740948 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-79kj9"] Dec 04 14:17:26 crc kubenswrapper[4715]: I1204 14:17:26.883509 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5r72w"] Dec 04 14:17:26 crc kubenswrapper[4715]: W1204 14:17:26.892951 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2386cbfc_2eba_4ace_a56a_0e203b7b9229.slice/crio-5c26d9ce6868acf2740fd8ad2296beca6fddce749821857bdc843e7e4c64a29a WatchSource:0}: Error finding container 5c26d9ce6868acf2740fd8ad2296beca6fddce749821857bdc843e7e4c64a29a: Status 404 returned error can't find the container with id 5c26d9ce6868acf2740fd8ad2296beca6fddce749821857bdc843e7e4c64a29a Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.156612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:27 crc kubenswrapper[4715]: E1204 14:17:27.156845 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:17:27 crc kubenswrapper[4715]: E1204 14:17:27.156863 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:17:27 crc kubenswrapper[4715]: E1204 14:17:27.156914 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift podName:548d91b6-2482-4624-843a-ea2898f7beb8 nodeName:}" failed. No retries permitted until 2025-12-04 14:17:29.156894759 +0000 UTC m=+1206.225612994 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift") pod "swift-storage-0" (UID: "548d91b6-2482-4624-843a-ea2898f7beb8") : configmap "swift-ring-files" not found Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.192581 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f92de975-e892-45d7-86b9-45dd9e4115b5" path="/var/lib/kubelet/pods/f92de975-e892-45d7-86b9-45dd9e4115b5/volumes" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.707459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5r72w" event={"ID":"2386cbfc-2eba-4ace-a56a-0e203b7b9229","Type":"ContainerStarted","Data":"5c26d9ce6868acf2740fd8ad2296beca6fddce749821857bdc843e7e4c64a29a"} Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.777505 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-2988-account-create-update-9spll"] Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.778654 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.782102 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.784614 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2988-account-create-update-9spll"] Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.831230 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-74hx8"] Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.834777 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74hx8" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.846327 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-74hx8"] Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.873953 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tfrz\" (UniqueName: \"kubernetes.io/projected/b8932cfe-78f8-4b68-9312-db06249648aa-kube-api-access-4tfrz\") pod \"glance-db-create-74hx8\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " pod="openstack/glance-db-create-74hx8" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.874054 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-operator-scripts\") pod \"glance-2988-account-create-update-9spll\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.874162 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vthzb\" (UniqueName: \"kubernetes.io/projected/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-kube-api-access-vthzb\") pod \"glance-2988-account-create-update-9spll\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.874228 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8932cfe-78f8-4b68-9312-db06249648aa-operator-scripts\") pod \"glance-db-create-74hx8\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " pod="openstack/glance-db-create-74hx8" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.981729 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8932cfe-78f8-4b68-9312-db06249648aa-operator-scripts\") pod \"glance-db-create-74hx8\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " pod="openstack/glance-db-create-74hx8" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.981852 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tfrz\" (UniqueName: \"kubernetes.io/projected/b8932cfe-78f8-4b68-9312-db06249648aa-kube-api-access-4tfrz\") pod \"glance-db-create-74hx8\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " pod="openstack/glance-db-create-74hx8" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.981894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-operator-scripts\") pod \"glance-2988-account-create-update-9spll\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.981937 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vthzb\" (UniqueName: \"kubernetes.io/projected/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-kube-api-access-vthzb\") pod \"glance-2988-account-create-update-9spll\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.982727 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8932cfe-78f8-4b68-9312-db06249648aa-operator-scripts\") pod \"glance-db-create-74hx8\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " pod="openstack/glance-db-create-74hx8" Dec 04 14:17:27 crc kubenswrapper[4715]: I1204 14:17:27.982860 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-operator-scripts\") pod \"glance-2988-account-create-update-9spll\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:28 crc kubenswrapper[4715]: I1204 14:17:28.005211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vthzb\" (UniqueName: \"kubernetes.io/projected/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-kube-api-access-vthzb\") pod \"glance-2988-account-create-update-9spll\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:28 crc kubenswrapper[4715]: I1204 14:17:28.018730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tfrz\" (UniqueName: \"kubernetes.io/projected/b8932cfe-78f8-4b68-9312-db06249648aa-kube-api-access-4tfrz\") pod \"glance-db-create-74hx8\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " pod="openstack/glance-db-create-74hx8" Dec 04 14:17:28 crc kubenswrapper[4715]: I1204 14:17:28.110692 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:28 crc kubenswrapper[4715]: I1204 14:17:28.150442 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74hx8" Dec 04 14:17:29 crc kubenswrapper[4715]: I1204 14:17:29.202601 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:29 crc kubenswrapper[4715]: E1204 14:17:29.203926 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:17:29 crc kubenswrapper[4715]: E1204 14:17:29.203942 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:17:29 crc kubenswrapper[4715]: E1204 14:17:29.204010 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift podName:548d91b6-2482-4624-843a-ea2898f7beb8 nodeName:}" failed. No retries permitted until 2025-12-04 14:17:33.203995344 +0000 UTC m=+1210.272713559 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift") pod "swift-storage-0" (UID: "548d91b6-2482-4624-843a-ea2898f7beb8") : configmap "swift-ring-files" not found Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.061562 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-sfl96"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.070078 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sfl96"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.070393 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.150693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f167ccc4-f804-4a6e-b194-0274a320e900-operator-scripts\") pod \"keystone-db-create-sfl96\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.150792 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmvjv\" (UniqueName: \"kubernetes.io/projected/f167ccc4-f804-4a6e-b194-0274a320e900-kube-api-access-hmvjv\") pod \"keystone-db-create-sfl96\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.158883 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-1742-account-create-update-gxqqt"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.169145 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.173772 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.194889 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1742-account-create-update-gxqqt"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.240676 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-74hx8"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.254166 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f167ccc4-f804-4a6e-b194-0274a320e900-operator-scripts\") pod \"keystone-db-create-sfl96\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.254224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmvjv\" (UniqueName: \"kubernetes.io/projected/f167ccc4-f804-4a6e-b194-0274a320e900-kube-api-access-hmvjv\") pod \"keystone-db-create-sfl96\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.255338 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f167ccc4-f804-4a6e-b194-0274a320e900-operator-scripts\") pod \"keystone-db-create-sfl96\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.271250 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmvjv\" (UniqueName: \"kubernetes.io/projected/f167ccc4-f804-4a6e-b194-0274a320e900-kube-api-access-hmvjv\") pod \"keystone-db-create-sfl96\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.355345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b5b100-cb32-4476-9631-74cb35856752-operator-scripts\") pod \"keystone-1742-account-create-update-gxqqt\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.355513 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtpm\" (UniqueName: \"kubernetes.io/projected/e7b5b100-cb32-4476-9631-74cb35856752-kube-api-access-vrtpm\") pod \"keystone-1742-account-create-update-gxqqt\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.357583 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-nc788"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.359132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.388997 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nc788"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.400599 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.410653 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2988-account-create-update-9spll"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.457174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtpm\" (UniqueName: \"kubernetes.io/projected/e7b5b100-cb32-4476-9631-74cb35856752-kube-api-access-vrtpm\") pod \"keystone-1742-account-create-update-gxqqt\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.457295 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b5b100-cb32-4476-9631-74cb35856752-operator-scripts\") pod \"keystone-1742-account-create-update-gxqqt\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.458075 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b5b100-cb32-4476-9631-74cb35856752-operator-scripts\") pod \"keystone-1742-account-create-update-gxqqt\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.468062 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-890e-account-create-update-q5wjc"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.469303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.476534 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.478600 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-890e-account-create-update-q5wjc"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.478724 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtpm\" (UniqueName: \"kubernetes.io/projected/e7b5b100-cb32-4476-9631-74cb35856752-kube-api-access-vrtpm\") pod \"keystone-1742-account-create-update-gxqqt\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.498826 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.562386 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-operator-scripts\") pod \"placement-db-create-nc788\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.562492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jnc6\" (UniqueName: \"kubernetes.io/projected/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-kube-api-access-9jnc6\") pod \"placement-db-create-nc788\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.663905 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcdhn\" (UniqueName: \"kubernetes.io/projected/99af1217-29e9-43e2-bb9c-441670a08c75-kube-api-access-tcdhn\") pod \"placement-890e-account-create-update-q5wjc\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.664042 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-operator-scripts\") pod \"placement-db-create-nc788\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.664076 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99af1217-29e9-43e2-bb9c-441670a08c75-operator-scripts\") pod \"placement-890e-account-create-update-q5wjc\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.664121 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jnc6\" (UniqueName: \"kubernetes.io/projected/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-kube-api-access-9jnc6\") pod \"placement-db-create-nc788\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.665061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-operator-scripts\") pod \"placement-db-create-nc788\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.687718 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jnc6\" (UniqueName: \"kubernetes.io/projected/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-kube-api-access-9jnc6\") pod \"placement-db-create-nc788\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " pod="openstack/placement-db-create-nc788" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.755804 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5r72w" event={"ID":"2386cbfc-2eba-4ace-a56a-0e203b7b9229","Type":"ContainerStarted","Data":"5cb48c6308513fed6fe7e7dc1387963b20173b8586224cd237ef1bdf78d37373"} Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.758029 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2988-account-create-update-9spll" event={"ID":"ab09ee34-7f1d-4a57-ac0d-0d809445ba95","Type":"ContainerStarted","Data":"fa9754f3f8d690c42db6b5416de8132665cc05d7891cfc44f59dafd7c02379c8"} Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.758085 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2988-account-create-update-9spll" event={"ID":"ab09ee34-7f1d-4a57-ac0d-0d809445ba95","Type":"ContainerStarted","Data":"3926c023160bef781eda8523794aecb46987981f8281cc388019fe380bd79b76"} Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.765288 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99af1217-29e9-43e2-bb9c-441670a08c75-operator-scripts\") pod \"placement-890e-account-create-update-q5wjc\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.766766 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcdhn\" (UniqueName: \"kubernetes.io/projected/99af1217-29e9-43e2-bb9c-441670a08c75-kube-api-access-tcdhn\") pod \"placement-890e-account-create-update-q5wjc\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.767262 4715 generic.go:334] "Generic (PLEG): container finished" podID="b8932cfe-78f8-4b68-9312-db06249648aa" containerID="71706ad94f1431d9a3543e36e9034bf7365e382f38af42bf3a493e38a2467ae5" exitCode=0 Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.767307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74hx8" event={"ID":"b8932cfe-78f8-4b68-9312-db06249648aa","Type":"ContainerDied","Data":"71706ad94f1431d9a3543e36e9034bf7365e382f38af42bf3a493e38a2467ae5"} Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.767333 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74hx8" event={"ID":"b8932cfe-78f8-4b68-9312-db06249648aa","Type":"ContainerStarted","Data":"9606eaedee79b2526873cd916e4548c719e1f51fbe46941f5cac6fbdf3215392"} Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.767601 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99af1217-29e9-43e2-bb9c-441670a08c75-operator-scripts\") pod \"placement-890e-account-create-update-q5wjc\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.771541 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sfl96"] Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.792349 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-5r72w" podStartSLOduration=1.863146098 podStartE2EDuration="6.792306061s" podCreationTimestamp="2025-12-04 14:17:26 +0000 UTC" firstStartedPulling="2025-12-04 14:17:26.895116596 +0000 UTC m=+1203.963834811" lastFinishedPulling="2025-12-04 14:17:31.824276549 +0000 UTC m=+1208.892994774" observedRunningTime="2025-12-04 14:17:32.777245459 +0000 UTC m=+1209.845963674" watchObservedRunningTime="2025-12-04 14:17:32.792306061 +0000 UTC m=+1209.861024276" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.793576 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcdhn\" (UniqueName: \"kubernetes.io/projected/99af1217-29e9-43e2-bb9c-441670a08c75-kube-api-access-tcdhn\") pod \"placement-890e-account-create-update-q5wjc\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.818815 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-2988-account-create-update-9spll" podStartSLOduration=5.818754246 podStartE2EDuration="5.818754246s" podCreationTimestamp="2025-12-04 14:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:32.797954152 +0000 UTC m=+1209.866672387" watchObservedRunningTime="2025-12-04 14:17:32.818754246 +0000 UTC m=+1209.887472471" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.846149 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:32 crc kubenswrapper[4715]: I1204 14:17:32.983818 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nc788" Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.074658 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1742-account-create-update-gxqqt"] Dec 04 14:17:33 crc kubenswrapper[4715]: W1204 14:17:33.076228 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7b5b100_cb32_4476_9631_74cb35856752.slice/crio-dd4014092b87574411cdbf4a0a1aea33ed627bf28f20081317d511c3c612b259 WatchSource:0}: Error finding container dd4014092b87574411cdbf4a0a1aea33ed627bf28f20081317d511c3c612b259: Status 404 returned error can't find the container with id dd4014092b87574411cdbf4a0a1aea33ed627bf28f20081317d511c3c612b259 Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.276651 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:33 crc kubenswrapper[4715]: E1204 14:17:33.277794 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:17:33 crc kubenswrapper[4715]: E1204 14:17:33.277882 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:17:33 crc kubenswrapper[4715]: E1204 14:17:33.277954 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift podName:548d91b6-2482-4624-843a-ea2898f7beb8 nodeName:}" failed. No retries permitted until 2025-12-04 14:17:41.277935425 +0000 UTC m=+1218.346653640 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift") pod "swift-storage-0" (UID: "548d91b6-2482-4624-843a-ea2898f7beb8") : configmap "swift-ring-files" not found Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.294011 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-890e-account-create-update-q5wjc"] Dec 04 14:17:33 crc kubenswrapper[4715]: W1204 14:17:33.295134 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99af1217_29e9_43e2_bb9c_441670a08c75.slice/crio-31c4f01f46baabf0a0ab72e114c7c8233227594f31ae364cd70f05f4a63f0dba WatchSource:0}: Error finding container 31c4f01f46baabf0a0ab72e114c7c8233227594f31ae364cd70f05f4a63f0dba: Status 404 returned error can't find the container with id 31c4f01f46baabf0a0ab72e114c7c8233227594f31ae364cd70f05f4a63f0dba Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.458086 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nc788"] Dec 04 14:17:33 crc kubenswrapper[4715]: W1204 14:17:33.463464 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf0daee2_9146_4c46_b3f9_8c4008eb6b7b.slice/crio-41675804bda5def972dd9b9a7fcafe59a75e684a07d283b51cd2c85fdc44c498 WatchSource:0}: Error finding container 41675804bda5def972dd9b9a7fcafe59a75e684a07d283b51cd2c85fdc44c498: Status 404 returned error can't find the container with id 41675804bda5def972dd9b9a7fcafe59a75e684a07d283b51cd2c85fdc44c498 Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.776890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nc788" event={"ID":"af0daee2-9146-4c46-b3f9-8c4008eb6b7b","Type":"ContainerStarted","Data":"41675804bda5def972dd9b9a7fcafe59a75e684a07d283b51cd2c85fdc44c498"} Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.779341 4715 generic.go:334] "Generic (PLEG): container finished" podID="ab09ee34-7f1d-4a57-ac0d-0d809445ba95" containerID="fa9754f3f8d690c42db6b5416de8132665cc05d7891cfc44f59dafd7c02379c8" exitCode=0 Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.779454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2988-account-create-update-9spll" event={"ID":"ab09ee34-7f1d-4a57-ac0d-0d809445ba95","Type":"ContainerDied","Data":"fa9754f3f8d690c42db6b5416de8132665cc05d7891cfc44f59dafd7c02379c8"} Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.781257 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sfl96" event={"ID":"f167ccc4-f804-4a6e-b194-0274a320e900","Type":"ContainerStarted","Data":"d0d60bf9da736590699534719c054268068a7018408053753ad7ccd4fed47ba0"} Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.781285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sfl96" event={"ID":"f167ccc4-f804-4a6e-b194-0274a320e900","Type":"ContainerStarted","Data":"805dc2ef93e5e8f1388d7b8902f7875313cd42bf72b588cd13e5cafea5621d6a"} Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.783025 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-890e-account-create-update-q5wjc" event={"ID":"99af1217-29e9-43e2-bb9c-441670a08c75","Type":"ContainerStarted","Data":"31c4f01f46baabf0a0ab72e114c7c8233227594f31ae364cd70f05f4a63f0dba"} Dec 04 14:17:33 crc kubenswrapper[4715]: I1204 14:17:33.790076 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1742-account-create-update-gxqqt" event={"ID":"e7b5b100-cb32-4476-9631-74cb35856752","Type":"ContainerStarted","Data":"dd4014092b87574411cdbf4a0a1aea33ed627bf28f20081317d511c3c612b259"} Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.179538 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74hx8" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.196609 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tfrz\" (UniqueName: \"kubernetes.io/projected/b8932cfe-78f8-4b68-9312-db06249648aa-kube-api-access-4tfrz\") pod \"b8932cfe-78f8-4b68-9312-db06249648aa\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.196856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8932cfe-78f8-4b68-9312-db06249648aa-operator-scripts\") pod \"b8932cfe-78f8-4b68-9312-db06249648aa\" (UID: \"b8932cfe-78f8-4b68-9312-db06249648aa\") " Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.198352 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8932cfe-78f8-4b68-9312-db06249648aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8932cfe-78f8-4b68-9312-db06249648aa" (UID: "b8932cfe-78f8-4b68-9312-db06249648aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.220544 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8932cfe-78f8-4b68-9312-db06249648aa-kube-api-access-4tfrz" (OuterVolumeSpecName: "kube-api-access-4tfrz") pod "b8932cfe-78f8-4b68-9312-db06249648aa" (UID: "b8932cfe-78f8-4b68-9312-db06249648aa"). InnerVolumeSpecName "kube-api-access-4tfrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.299131 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tfrz\" (UniqueName: \"kubernetes.io/projected/b8932cfe-78f8-4b68-9312-db06249648aa-kube-api-access-4tfrz\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.299169 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8932cfe-78f8-4b68-9312-db06249648aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.580250 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.640842 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5c45t"] Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.641132 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerName="dnsmasq-dns" containerID="cri-o://3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f" gracePeriod=10 Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.800021 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-890e-account-create-update-q5wjc" event={"ID":"99af1217-29e9-43e2-bb9c-441670a08c75","Type":"ContainerStarted","Data":"ff443cd073060401f9a7bb24699b7459b6f14d7447e578263aef790f1d3ff8aa"} Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.805539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1742-account-create-update-gxqqt" event={"ID":"e7b5b100-cb32-4476-9631-74cb35856752","Type":"ContainerStarted","Data":"0accca53065ad2257f5d65487256dc53eedeb76d65ce657e0d238469c9fb2b4d"} Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.814371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nc788" event={"ID":"af0daee2-9146-4c46-b3f9-8c4008eb6b7b","Type":"ContainerStarted","Data":"243b1152e8f12ffcdbfa588c944c85c6389fdb18a26490bb9ca63e2b14687750"} Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.823857 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-890e-account-create-update-q5wjc" podStartSLOduration=2.823835512 podStartE2EDuration="2.823835512s" podCreationTimestamp="2025-12-04 14:17:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:34.819501066 +0000 UTC m=+1211.888219291" watchObservedRunningTime="2025-12-04 14:17:34.823835512 +0000 UTC m=+1211.892553737" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.827687 4715 generic.go:334] "Generic (PLEG): container finished" podID="f167ccc4-f804-4a6e-b194-0274a320e900" containerID="d0d60bf9da736590699534719c054268068a7018408053753ad7ccd4fed47ba0" exitCode=0 Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.827864 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sfl96" event={"ID":"f167ccc4-f804-4a6e-b194-0274a320e900","Type":"ContainerDied","Data":"d0d60bf9da736590699534719c054268068a7018408053753ad7ccd4fed47ba0"} Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.830752 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74hx8" event={"ID":"b8932cfe-78f8-4b68-9312-db06249648aa","Type":"ContainerDied","Data":"9606eaedee79b2526873cd916e4548c719e1f51fbe46941f5cac6fbdf3215392"} Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.830816 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9606eaedee79b2526873cd916e4548c719e1f51fbe46941f5cac6fbdf3215392" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.830770 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74hx8" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.849964 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-1742-account-create-update-gxqqt" podStartSLOduration=2.849940738 podStartE2EDuration="2.849940738s" podCreationTimestamp="2025-12-04 14:17:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:34.832183434 +0000 UTC m=+1211.900901649" watchObservedRunningTime="2025-12-04 14:17:34.849940738 +0000 UTC m=+1211.918658963" Dec 04 14:17:34 crc kubenswrapper[4715]: I1204 14:17:34.852738 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-nc788" podStartSLOduration=2.852724792 podStartE2EDuration="2.852724792s" podCreationTimestamp="2025-12-04 14:17:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:34.847210585 +0000 UTC m=+1211.915928820" watchObservedRunningTime="2025-12-04 14:17:34.852724792 +0000 UTC m=+1211.921443017" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.249269 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.448154 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.535273 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vthzb\" (UniqueName: \"kubernetes.io/projected/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-kube-api-access-vthzb\") pod \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.535542 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-operator-scripts\") pod \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\" (UID: \"ab09ee34-7f1d-4a57-ac0d-0d809445ba95\") " Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.535973 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab09ee34-7f1d-4a57-ac0d-0d809445ba95" (UID: "ab09ee34-7f1d-4a57-ac0d-0d809445ba95"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.544367 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-kube-api-access-vthzb" (OuterVolumeSpecName: "kube-api-access-vthzb") pod "ab09ee34-7f1d-4a57-ac0d-0d809445ba95" (UID: "ab09ee34-7f1d-4a57-ac0d-0d809445ba95"). InnerVolumeSpecName "kube-api-access-vthzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.640711 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.640745 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vthzb\" (UniqueName: \"kubernetes.io/projected/ab09ee34-7f1d-4a57-ac0d-0d809445ba95-kube-api-access-vthzb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.644536 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.741483 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgdww\" (UniqueName: \"kubernetes.io/projected/6d6533ec-6214-4aa6-828c-707f1d9f8f11-kube-api-access-kgdww\") pod \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.741597 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-dns-svc\") pod \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.741664 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-config\") pod \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\" (UID: \"6d6533ec-6214-4aa6-828c-707f1d9f8f11\") " Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.744501 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d6533ec-6214-4aa6-828c-707f1d9f8f11-kube-api-access-kgdww" (OuterVolumeSpecName: "kube-api-access-kgdww") pod "6d6533ec-6214-4aa6-828c-707f1d9f8f11" (UID: "6d6533ec-6214-4aa6-828c-707f1d9f8f11"). InnerVolumeSpecName "kube-api-access-kgdww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.779944 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-config" (OuterVolumeSpecName: "config") pod "6d6533ec-6214-4aa6-828c-707f1d9f8f11" (UID: "6d6533ec-6214-4aa6-828c-707f1d9f8f11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.781428 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6d6533ec-6214-4aa6-828c-707f1d9f8f11" (UID: "6d6533ec-6214-4aa6-828c-707f1d9f8f11"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.841792 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerID="3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f" exitCode=0 Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.841858 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.841868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" event={"ID":"6d6533ec-6214-4aa6-828c-707f1d9f8f11","Type":"ContainerDied","Data":"3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.841900 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5c45t" event={"ID":"6d6533ec-6214-4aa6-828c-707f1d9f8f11","Type":"ContainerDied","Data":"3dd80161d389dded94aa3fa48c59e54a87f47d688d8e908a17ecf85c3c282de5"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.841919 4715 scope.go:117] "RemoveContainer" containerID="3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.842851 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.842868 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgdww\" (UniqueName: \"kubernetes.io/projected/6d6533ec-6214-4aa6-828c-707f1d9f8f11-kube-api-access-kgdww\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.842878 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d6533ec-6214-4aa6-828c-707f1d9f8f11-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.843443 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7b5b100-cb32-4476-9631-74cb35856752" containerID="0accca53065ad2257f5d65487256dc53eedeb76d65ce657e0d238469c9fb2b4d" exitCode=0 Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.843506 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1742-account-create-update-gxqqt" event={"ID":"e7b5b100-cb32-4476-9631-74cb35856752","Type":"ContainerDied","Data":"0accca53065ad2257f5d65487256dc53eedeb76d65ce657e0d238469c9fb2b4d"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.845677 4715 generic.go:334] "Generic (PLEG): container finished" podID="af0daee2-9146-4c46-b3f9-8c4008eb6b7b" containerID="243b1152e8f12ffcdbfa588c944c85c6389fdb18a26490bb9ca63e2b14687750" exitCode=0 Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.845750 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nc788" event={"ID":"af0daee2-9146-4c46-b3f9-8c4008eb6b7b","Type":"ContainerDied","Data":"243b1152e8f12ffcdbfa588c944c85c6389fdb18a26490bb9ca63e2b14687750"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.847460 4715 generic.go:334] "Generic (PLEG): container finished" podID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerID="70e712e943ad0c774cffe640698edc91acb2e0c08db4632f0be92d036d4b9251" exitCode=0 Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.847498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"440f7f8d-a89d-4d6d-8280-c56fede1c91c","Type":"ContainerDied","Data":"70e712e943ad0c774cffe640698edc91acb2e0c08db4632f0be92d036d4b9251"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.849686 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2988-account-create-update-9spll" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.849690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2988-account-create-update-9spll" event={"ID":"ab09ee34-7f1d-4a57-ac0d-0d809445ba95","Type":"ContainerDied","Data":"3926c023160bef781eda8523794aecb46987981f8281cc388019fe380bd79b76"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.849739 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3926c023160bef781eda8523794aecb46987981f8281cc388019fe380bd79b76" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.851589 4715 generic.go:334] "Generic (PLEG): container finished" podID="99af1217-29e9-43e2-bb9c-441670a08c75" containerID="ff443cd073060401f9a7bb24699b7459b6f14d7447e578263aef790f1d3ff8aa" exitCode=0 Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.852249 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-890e-account-create-update-q5wjc" event={"ID":"99af1217-29e9-43e2-bb9c-441670a08c75","Type":"ContainerDied","Data":"ff443cd073060401f9a7bb24699b7459b6f14d7447e578263aef790f1d3ff8aa"} Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.883078 4715 scope.go:117] "RemoveContainer" containerID="580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.889975 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5c45t"] Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.902425 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5c45t"] Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.908631 4715 scope.go:117] "RemoveContainer" containerID="3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f" Dec 04 14:17:35 crc kubenswrapper[4715]: E1204 14:17:35.909584 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f\": container with ID starting with 3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f not found: ID does not exist" containerID="3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.909662 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f"} err="failed to get container status \"3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f\": rpc error: code = NotFound desc = could not find container \"3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f\": container with ID starting with 3f6156579fc957e0ce7c6aa62881026ee74f6edd33905c30f3fcaaf01ff4090f not found: ID does not exist" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.909696 4715 scope.go:117] "RemoveContainer" containerID="580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748" Dec 04 14:17:35 crc kubenswrapper[4715]: E1204 14:17:35.910076 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748\": container with ID starting with 580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748 not found: ID does not exist" containerID="580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748" Dec 04 14:17:35 crc kubenswrapper[4715]: I1204 14:17:35.910192 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748"} err="failed to get container status \"580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748\": rpc error: code = NotFound desc = could not find container \"580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748\": container with ID starting with 580245695db2efbcb112998c6ac6d2f1b5683d3592d1d6c824c2ea59f13b3748 not found: ID does not exist" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.244169 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.351917 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f167ccc4-f804-4a6e-b194-0274a320e900-operator-scripts\") pod \"f167ccc4-f804-4a6e-b194-0274a320e900\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.352156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmvjv\" (UniqueName: \"kubernetes.io/projected/f167ccc4-f804-4a6e-b194-0274a320e900-kube-api-access-hmvjv\") pod \"f167ccc4-f804-4a6e-b194-0274a320e900\" (UID: \"f167ccc4-f804-4a6e-b194-0274a320e900\") " Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.352725 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f167ccc4-f804-4a6e-b194-0274a320e900-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f167ccc4-f804-4a6e-b194-0274a320e900" (UID: "f167ccc4-f804-4a6e-b194-0274a320e900"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.357248 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f167ccc4-f804-4a6e-b194-0274a320e900-kube-api-access-hmvjv" (OuterVolumeSpecName: "kube-api-access-hmvjv") pod "f167ccc4-f804-4a6e-b194-0274a320e900" (UID: "f167ccc4-f804-4a6e-b194-0274a320e900"). InnerVolumeSpecName "kube-api-access-hmvjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.454192 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmvjv\" (UniqueName: \"kubernetes.io/projected/f167ccc4-f804-4a6e-b194-0274a320e900-kube-api-access-hmvjv\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.454225 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f167ccc4-f804-4a6e-b194-0274a320e900-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.862699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sfl96" event={"ID":"f167ccc4-f804-4a6e-b194-0274a320e900","Type":"ContainerDied","Data":"805dc2ef93e5e8f1388d7b8902f7875313cd42bf72b588cd13e5cafea5621d6a"} Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.862739 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="805dc2ef93e5e8f1388d7b8902f7875313cd42bf72b588cd13e5cafea5621d6a" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.862805 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sfl96" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.865371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"440f7f8d-a89d-4d6d-8280-c56fede1c91c","Type":"ContainerStarted","Data":"8ab6c67abc9e697ad6cc5592319e9baccda512a795b344eb06ca040df3be5cd9"} Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.865826 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 14:17:36 crc kubenswrapper[4715]: I1204 14:17:36.898343 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.760460825 podStartE2EDuration="59.898323907s" podCreationTimestamp="2025-12-04 14:16:37 +0000 UTC" firstStartedPulling="2025-12-04 14:16:39.976063663 +0000 UTC m=+1157.044781878" lastFinishedPulling="2025-12-04 14:17:01.113926735 +0000 UTC m=+1178.182644960" observedRunningTime="2025-12-04 14:17:36.891538106 +0000 UTC m=+1213.960256321" watchObservedRunningTime="2025-12-04 14:17:36.898323907 +0000 UTC m=+1213.967042122" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.191419 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" path="/var/lib/kubelet/pods/6d6533ec-6214-4aa6-828c-707f1d9f8f11/volumes" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.373783 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.382577 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.392972 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nc788" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574015 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jnc6\" (UniqueName: \"kubernetes.io/projected/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-kube-api-access-9jnc6\") pod \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574099 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrtpm\" (UniqueName: \"kubernetes.io/projected/e7b5b100-cb32-4476-9631-74cb35856752-kube-api-access-vrtpm\") pod \"e7b5b100-cb32-4476-9631-74cb35856752\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574137 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcdhn\" (UniqueName: \"kubernetes.io/projected/99af1217-29e9-43e2-bb9c-441670a08c75-kube-api-access-tcdhn\") pod \"99af1217-29e9-43e2-bb9c-441670a08c75\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b5b100-cb32-4476-9631-74cb35856752-operator-scripts\") pod \"e7b5b100-cb32-4476-9631-74cb35856752\" (UID: \"e7b5b100-cb32-4476-9631-74cb35856752\") " Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574334 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99af1217-29e9-43e2-bb9c-441670a08c75-operator-scripts\") pod \"99af1217-29e9-43e2-bb9c-441670a08c75\" (UID: \"99af1217-29e9-43e2-bb9c-441670a08c75\") " Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574440 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-operator-scripts\") pod \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\" (UID: \"af0daee2-9146-4c46-b3f9-8c4008eb6b7b\") " Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.574701 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7b5b100-cb32-4476-9631-74cb35856752-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7b5b100-cb32-4476-9631-74cb35856752" (UID: "e7b5b100-cb32-4476-9631-74cb35856752"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.575082 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af0daee2-9146-4c46-b3f9-8c4008eb6b7b" (UID: "af0daee2-9146-4c46-b3f9-8c4008eb6b7b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.575108 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99af1217-29e9-43e2-bb9c-441670a08c75-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99af1217-29e9-43e2-bb9c-441670a08c75" (UID: "99af1217-29e9-43e2-bb9c-441670a08c75"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.575448 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b5b100-cb32-4476-9631-74cb35856752-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.575484 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99af1217-29e9-43e2-bb9c-441670a08c75-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.575496 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.579506 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-kube-api-access-9jnc6" (OuterVolumeSpecName: "kube-api-access-9jnc6") pod "af0daee2-9146-4c46-b3f9-8c4008eb6b7b" (UID: "af0daee2-9146-4c46-b3f9-8c4008eb6b7b"). InnerVolumeSpecName "kube-api-access-9jnc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.580235 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99af1217-29e9-43e2-bb9c-441670a08c75-kube-api-access-tcdhn" (OuterVolumeSpecName: "kube-api-access-tcdhn") pod "99af1217-29e9-43e2-bb9c-441670a08c75" (UID: "99af1217-29e9-43e2-bb9c-441670a08c75"). InnerVolumeSpecName "kube-api-access-tcdhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.581259 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b5b100-cb32-4476-9631-74cb35856752-kube-api-access-vrtpm" (OuterVolumeSpecName: "kube-api-access-vrtpm") pod "e7b5b100-cb32-4476-9631-74cb35856752" (UID: "e7b5b100-cb32-4476-9631-74cb35856752"). InnerVolumeSpecName "kube-api-access-vrtpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.676850 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcdhn\" (UniqueName: \"kubernetes.io/projected/99af1217-29e9-43e2-bb9c-441670a08c75-kube-api-access-tcdhn\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.677495 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jnc6\" (UniqueName: \"kubernetes.io/projected/af0daee2-9146-4c46-b3f9-8c4008eb6b7b-kube-api-access-9jnc6\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.677528 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrtpm\" (UniqueName: \"kubernetes.io/projected/e7b5b100-cb32-4476-9631-74cb35856752-kube-api-access-vrtpm\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.876233 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1742-account-create-update-gxqqt" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.876222 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1742-account-create-update-gxqqt" event={"ID":"e7b5b100-cb32-4476-9631-74cb35856752","Type":"ContainerDied","Data":"dd4014092b87574411cdbf4a0a1aea33ed627bf28f20081317d511c3c612b259"} Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.876391 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd4014092b87574411cdbf4a0a1aea33ed627bf28f20081317d511c3c612b259" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.878654 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nc788" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.878955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nc788" event={"ID":"af0daee2-9146-4c46-b3f9-8c4008eb6b7b","Type":"ContainerDied","Data":"41675804bda5def972dd9b9a7fcafe59a75e684a07d283b51cd2c85fdc44c498"} Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.879072 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41675804bda5def972dd9b9a7fcafe59a75e684a07d283b51cd2c85fdc44c498" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.881854 4715 generic.go:334] "Generic (PLEG): container finished" podID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerID="41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e" exitCode=0 Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.881947 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34885f9-cf98-4b1a-bbf3-c7af62f5f273","Type":"ContainerDied","Data":"41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e"} Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.885082 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-890e-account-create-update-q5wjc" Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.885605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-890e-account-create-update-q5wjc" event={"ID":"99af1217-29e9-43e2-bb9c-441670a08c75","Type":"ContainerDied","Data":"31c4f01f46baabf0a0ab72e114c7c8233227594f31ae364cd70f05f4a63f0dba"} Dec 04 14:17:37 crc kubenswrapper[4715]: I1204 14:17:37.885717 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c4f01f46baabf0a0ab72e114c7c8233227594f31ae364cd70f05f4a63f0dba" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039223 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ddj77"] Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039853 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerName="dnsmasq-dns" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039870 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerName="dnsmasq-dns" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039882 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b5b100-cb32-4476-9631-74cb35856752" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039888 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b5b100-cb32-4476-9631-74cb35856752" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039899 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerName="init" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039905 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerName="init" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039914 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99af1217-29e9-43e2-bb9c-441670a08c75" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039921 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="99af1217-29e9-43e2-bb9c-441670a08c75" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039942 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab09ee34-7f1d-4a57-ac0d-0d809445ba95" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039948 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab09ee34-7f1d-4a57-ac0d-0d809445ba95" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039959 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0daee2-9146-4c46-b3f9-8c4008eb6b7b" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039967 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0daee2-9146-4c46-b3f9-8c4008eb6b7b" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039974 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8932cfe-78f8-4b68-9312-db06249648aa" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039981 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8932cfe-78f8-4b68-9312-db06249648aa" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: E1204 14:17:38.039993 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f167ccc4-f804-4a6e-b194-0274a320e900" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.039999 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f167ccc4-f804-4a6e-b194-0274a320e900" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040192 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6533ec-6214-4aa6-828c-707f1d9f8f11" containerName="dnsmasq-dns" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040203 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="99af1217-29e9-43e2-bb9c-441670a08c75" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040220 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0daee2-9146-4c46-b3f9-8c4008eb6b7b" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040231 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8932cfe-78f8-4b68-9312-db06249648aa" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040240 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f167ccc4-f804-4a6e-b194-0274a320e900" containerName="mariadb-database-create" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040252 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b5b100-cb32-4476-9631-74cb35856752" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040260 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab09ee34-7f1d-4a57-ac0d-0d809445ba95" containerName="mariadb-account-create-update" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.040814 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.043769 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.045894 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-886p6" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.055671 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ddj77"] Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.194615 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-combined-ca-bundle\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.194678 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-db-sync-config-data\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.194720 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c94g\" (UniqueName: \"kubernetes.io/projected/88b59e0b-f6fd-42e0-9780-7c13708c379c-kube-api-access-5c94g\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.194945 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-config-data\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.296660 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-combined-ca-bundle\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.296724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-db-sync-config-data\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.296766 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c94g\" (UniqueName: \"kubernetes.io/projected/88b59e0b-f6fd-42e0-9780-7c13708c379c-kube-api-access-5c94g\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.297224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-config-data\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.301257 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-combined-ca-bundle\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.302743 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-config-data\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.304257 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-db-sync-config-data\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.317924 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c94g\" (UniqueName: \"kubernetes.io/projected/88b59e0b-f6fd-42e0-9780-7c13708c379c-kube-api-access-5c94g\") pod \"glance-db-sync-ddj77\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.400199 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ddj77" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.758292 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.758660 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.893396 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34885f9-cf98-4b1a-bbf3-c7af62f5f273","Type":"ContainerStarted","Data":"6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71"} Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.894572 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.921720 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371974.933073 podStartE2EDuration="1m1.92170361s" podCreationTimestamp="2025-12-04 14:16:37 +0000 UTC" firstStartedPulling="2025-12-04 14:16:39.617658443 +0000 UTC m=+1156.686376658" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:38.920555059 +0000 UTC m=+1215.989273294" watchObservedRunningTime="2025-12-04 14:17:38.92170361 +0000 UTC m=+1215.990421825" Dec 04 14:17:38 crc kubenswrapper[4715]: I1204 14:17:38.952799 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ddj77"] Dec 04 14:17:38 crc kubenswrapper[4715]: W1204 14:17:38.963439 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88b59e0b_f6fd_42e0_9780_7c13708c379c.slice/crio-280c6bb2a0e2cf2e2aac764992e2dc20ea4ccdc46ce3250eb5ccdc199b561984 WatchSource:0}: Error finding container 280c6bb2a0e2cf2e2aac764992e2dc20ea4ccdc46ce3250eb5ccdc199b561984: Status 404 returned error can't find the container with id 280c6bb2a0e2cf2e2aac764992e2dc20ea4ccdc46ce3250eb5ccdc199b561984 Dec 04 14:17:39 crc kubenswrapper[4715]: I1204 14:17:39.906920 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ddj77" event={"ID":"88b59e0b-f6fd-42e0-9780-7c13708c379c","Type":"ContainerStarted","Data":"280c6bb2a0e2cf2e2aac764992e2dc20ea4ccdc46ce3250eb5ccdc199b561984"} Dec 04 14:17:41 crc kubenswrapper[4715]: I1204 14:17:41.377459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:41 crc kubenswrapper[4715]: E1204 14:17:41.377978 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:17:41 crc kubenswrapper[4715]: E1204 14:17:41.378060 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:17:41 crc kubenswrapper[4715]: E1204 14:17:41.378162 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift podName:548d91b6-2482-4624-843a-ea2898f7beb8 nodeName:}" failed. No retries permitted until 2025-12-04 14:17:57.378122804 +0000 UTC m=+1234.446841059 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift") pod "swift-storage-0" (UID: "548d91b6-2482-4624-843a-ea2898f7beb8") : configmap "swift-ring-files" not found Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.020776 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6k4jg" podUID="00743321-f084-4e95-b47d-1b21332be3d5" containerName="ovn-controller" probeResult="failure" output=< Dec 04 14:17:44 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 14:17:44 crc kubenswrapper[4715]: > Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.115846 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.133781 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vmdp8" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.380233 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6k4jg-config-689hp"] Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.381717 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.388956 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k4jg-config-689hp"] Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.395397 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.537102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-additional-scripts\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.537150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.537186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-log-ovn\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.537210 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9jz7\" (UniqueName: \"kubernetes.io/projected/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-kube-api-access-r9jz7\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.537257 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run-ovn\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.537292 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-scripts\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.638993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-log-ovn\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.639694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-log-ovn\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.639760 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9jz7\" (UniqueName: \"kubernetes.io/projected/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-kube-api-access-r9jz7\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.639932 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run-ovn\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.640023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-scripts\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.640198 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-additional-scripts\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.640227 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run-ovn\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.640241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.640320 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.641135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-additional-scripts\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.642602 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-scripts\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.660531 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9jz7\" (UniqueName: \"kubernetes.io/projected/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-kube-api-access-r9jz7\") pod \"ovn-controller-6k4jg-config-689hp\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:44 crc kubenswrapper[4715]: I1204 14:17:44.702338 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:45 crc kubenswrapper[4715]: I1204 14:17:45.384004 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k4jg-config-689hp"] Dec 04 14:17:46 crc kubenswrapper[4715]: I1204 14:17:46.017579 4715 generic.go:334] "Generic (PLEG): container finished" podID="2386cbfc-2eba-4ace-a56a-0e203b7b9229" containerID="5cb48c6308513fed6fe7e7dc1387963b20173b8586224cd237ef1bdf78d37373" exitCode=0 Dec 04 14:17:46 crc kubenswrapper[4715]: I1204 14:17:46.017628 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5r72w" event={"ID":"2386cbfc-2eba-4ace-a56a-0e203b7b9229","Type":"ContainerDied","Data":"5cb48c6308513fed6fe7e7dc1387963b20173b8586224cd237ef1bdf78d37373"} Dec 04 14:17:49 crc kubenswrapper[4715]: I1204 14:17:49.022267 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6k4jg" podUID="00743321-f084-4e95-b47d-1b21332be3d5" containerName="ovn-controller" probeResult="failure" output=< Dec 04 14:17:49 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 14:17:49 crc kubenswrapper[4715]: > Dec 04 14:17:49 crc kubenswrapper[4715]: I1204 14:17:49.067994 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 04 14:17:49 crc kubenswrapper[4715]: I1204 14:17:49.378913 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.018801 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6k4jg" podUID="00743321-f084-4e95-b47d-1b21332be3d5" containerName="ovn-controller" probeResult="failure" output=< Dec 04 14:17:54 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 14:17:54 crc kubenswrapper[4715]: > Dec 04 14:17:54 crc kubenswrapper[4715]: E1204 14:17:54.532613 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 04 14:17:54 crc kubenswrapper[4715]: E1204 14:17:54.533313 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5c94g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-ddj77_openstack(88b59e0b-f6fd-42e0-9780-7c13708c379c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:17:54 crc kubenswrapper[4715]: E1204 14:17:54.534965 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-ddj77" podUID="88b59e0b-f6fd-42e0-9780-7c13708c379c" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.607869 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.741858 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d46h\" (UniqueName: \"kubernetes.io/projected/2386cbfc-2eba-4ace-a56a-0e203b7b9229-kube-api-access-6d46h\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.742272 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-ring-data-devices\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.742312 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-scripts\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.742447 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-swiftconf\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.742487 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-combined-ca-bundle\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.742557 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2386cbfc-2eba-4ace-a56a-0e203b7b9229-etc-swift\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.742592 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-dispersionconf\") pod \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\" (UID: \"2386cbfc-2eba-4ace-a56a-0e203b7b9229\") " Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.743055 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.744324 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2386cbfc-2eba-4ace-a56a-0e203b7b9229-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.749190 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2386cbfc-2eba-4ace-a56a-0e203b7b9229-kube-api-access-6d46h" (OuterVolumeSpecName: "kube-api-access-6d46h") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "kube-api-access-6d46h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.753916 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.773599 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-scripts" (OuterVolumeSpecName: "scripts") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.776081 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.779304 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2386cbfc-2eba-4ace-a56a-0e203b7b9229" (UID: "2386cbfc-2eba-4ace-a56a-0e203b7b9229"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844388 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844423 4715 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2386cbfc-2eba-4ace-a56a-0e203b7b9229-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844432 4715 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844443 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d46h\" (UniqueName: \"kubernetes.io/projected/2386cbfc-2eba-4ace-a56a-0e203b7b9229-kube-api-access-6d46h\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844454 4715 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844463 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2386cbfc-2eba-4ace-a56a-0e203b7b9229-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:54 crc kubenswrapper[4715]: I1204 14:17:54.844472 4715 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2386cbfc-2eba-4ace-a56a-0e203b7b9229-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:55 crc kubenswrapper[4715]: I1204 14:17:55.102843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k4jg-config-689hp" event={"ID":"868a1540-37ed-4f8c-8a0b-2d44f99ab78b","Type":"ContainerStarted","Data":"ab30225d43018b007e016d184e5428ecce321a57c30879f40b3b33c0446f54cd"} Dec 04 14:17:55 crc kubenswrapper[4715]: I1204 14:17:55.102953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k4jg-config-689hp" event={"ID":"868a1540-37ed-4f8c-8a0b-2d44f99ab78b","Type":"ContainerStarted","Data":"b57d1231e0a658e7daf9e1d7dfae8adabe7b7fc2d5b5bb6956e4c216095fa0d5"} Dec 04 14:17:55 crc kubenswrapper[4715]: I1204 14:17:55.105928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5r72w" event={"ID":"2386cbfc-2eba-4ace-a56a-0e203b7b9229","Type":"ContainerDied","Data":"5c26d9ce6868acf2740fd8ad2296beca6fddce749821857bdc843e7e4c64a29a"} Dec 04 14:17:55 crc kubenswrapper[4715]: I1204 14:17:55.105996 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c26d9ce6868acf2740fd8ad2296beca6fddce749821857bdc843e7e4c64a29a" Dec 04 14:17:55 crc kubenswrapper[4715]: I1204 14:17:55.105941 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5r72w" Dec 04 14:17:55 crc kubenswrapper[4715]: E1204 14:17:55.107544 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-ddj77" podUID="88b59e0b-f6fd-42e0-9780-7c13708c379c" Dec 04 14:17:55 crc kubenswrapper[4715]: I1204 14:17:55.133292 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6k4jg-config-689hp" podStartSLOduration=11.133268845 podStartE2EDuration="11.133268845s" podCreationTimestamp="2025-12-04 14:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:55.122136798 +0000 UTC m=+1232.190855033" watchObservedRunningTime="2025-12-04 14:17:55.133268845 +0000 UTC m=+1232.201987060" Dec 04 14:17:56 crc kubenswrapper[4715]: I1204 14:17:56.113530 4715 generic.go:334] "Generic (PLEG): container finished" podID="868a1540-37ed-4f8c-8a0b-2d44f99ab78b" containerID="ab30225d43018b007e016d184e5428ecce321a57c30879f40b3b33c0446f54cd" exitCode=0 Dec 04 14:17:56 crc kubenswrapper[4715]: I1204 14:17:56.113592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k4jg-config-689hp" event={"ID":"868a1540-37ed-4f8c-8a0b-2d44f99ab78b","Type":"ContainerDied","Data":"ab30225d43018b007e016d184e5428ecce321a57c30879f40b3b33c0446f54cd"} Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.388441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.396449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/548d91b6-2482-4624-843a-ea2898f7beb8-etc-swift\") pod \"swift-storage-0\" (UID: \"548d91b6-2482-4624-843a-ea2898f7beb8\") " pod="openstack/swift-storage-0" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.463383 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.571128 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593462 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run-ovn\") pod \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593550 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-additional-scripts\") pod \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593581 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "868a1540-37ed-4f8c-8a0b-2d44f99ab78b" (UID: "868a1540-37ed-4f8c-8a0b-2d44f99ab78b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593599 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-log-ovn\") pod \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593631 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "868a1540-37ed-4f8c-8a0b-2d44f99ab78b" (UID: "868a1540-37ed-4f8c-8a0b-2d44f99ab78b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593722 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run\") pod \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593800 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run" (OuterVolumeSpecName: "var-run") pod "868a1540-37ed-4f8c-8a0b-2d44f99ab78b" (UID: "868a1540-37ed-4f8c-8a0b-2d44f99ab78b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593819 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-scripts\") pod \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.593863 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9jz7\" (UniqueName: \"kubernetes.io/projected/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-kube-api-access-r9jz7\") pod \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\" (UID: \"868a1540-37ed-4f8c-8a0b-2d44f99ab78b\") " Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.594264 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "868a1540-37ed-4f8c-8a0b-2d44f99ab78b" (UID: "868a1540-37ed-4f8c-8a0b-2d44f99ab78b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.594312 4715 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.594330 4715 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.594338 4715 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.594486 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-scripts" (OuterVolumeSpecName: "scripts") pod "868a1540-37ed-4f8c-8a0b-2d44f99ab78b" (UID: "868a1540-37ed-4f8c-8a0b-2d44f99ab78b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.601298 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-kube-api-access-r9jz7" (OuterVolumeSpecName: "kube-api-access-r9jz7") pod "868a1540-37ed-4f8c-8a0b-2d44f99ab78b" (UID: "868a1540-37ed-4f8c-8a0b-2d44f99ab78b"). InnerVolumeSpecName "kube-api-access-r9jz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.695956 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.696221 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9jz7\" (UniqueName: \"kubernetes.io/projected/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-kube-api-access-r9jz7\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:57 crc kubenswrapper[4715]: I1204 14:17:57.696234 4715 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/868a1540-37ed-4f8c-8a0b-2d44f99ab78b-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:58 crc kubenswrapper[4715]: I1204 14:17:58.116332 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 14:17:58 crc kubenswrapper[4715]: W1204 14:17:58.122929 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod548d91b6_2482_4624_843a_ea2898f7beb8.slice/crio-23dcb2d5f6404ede5c25eda5e4c5b76357cd066604526b011c8bb8375d745725 WatchSource:0}: Error finding container 23dcb2d5f6404ede5c25eda5e4c5b76357cd066604526b011c8bb8375d745725: Status 404 returned error can't find the container with id 23dcb2d5f6404ede5c25eda5e4c5b76357cd066604526b011c8bb8375d745725 Dec 04 14:17:58 crc kubenswrapper[4715]: I1204 14:17:58.129858 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k4jg-config-689hp" event={"ID":"868a1540-37ed-4f8c-8a0b-2d44f99ab78b","Type":"ContainerDied","Data":"b57d1231e0a658e7daf9e1d7dfae8adabe7b7fc2d5b5bb6956e4c216095fa0d5"} Dec 04 14:17:58 crc kubenswrapper[4715]: I1204 14:17:58.129909 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b57d1231e0a658e7daf9e1d7dfae8adabe7b7fc2d5b5bb6956e4c216095fa0d5" Dec 04 14:17:58 crc kubenswrapper[4715]: I1204 14:17:58.129979 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k4jg-config-689hp" Dec 04 14:17:58 crc kubenswrapper[4715]: I1204 14:17:58.231002 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6k4jg-config-689hp"] Dec 04 14:17:58 crc kubenswrapper[4715]: I1204 14:17:58.240687 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6k4jg-config-689hp"] Dec 04 14:17:59 crc kubenswrapper[4715]: I1204 14:17:59.014957 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6k4jg" Dec 04 14:17:59 crc kubenswrapper[4715]: I1204 14:17:59.072346 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:59 crc kubenswrapper[4715]: I1204 14:17:59.150067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"23dcb2d5f6404ede5c25eda5e4c5b76357cd066604526b011c8bb8375d745725"} Dec 04 14:17:59 crc kubenswrapper[4715]: I1204 14:17:59.198791 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="868a1540-37ed-4f8c-8a0b-2d44f99ab78b" path="/var/lib/kubelet/pods/868a1540-37ed-4f8c-8a0b-2d44f99ab78b/volumes" Dec 04 14:17:59 crc kubenswrapper[4715]: I1204 14:17:59.379340 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 14:18:00 crc kubenswrapper[4715]: I1204 14:18:00.161573 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"6c7224b22941debcd8174e45af31015c8a0c2d11e545192ed8d268f097727fa7"} Dec 04 14:18:00 crc kubenswrapper[4715]: I1204 14:18:00.162917 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"d5320c8a055728a86bd950d9ba1d457a88a58d2eaa1421a1af646822e5897133"} Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.178771 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"53e9f735732df0c779ab9b293fe9d543e1dd9620c4f6c1a01312d5e5ddfb8bdc"} Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.179127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"da5e93881e1e566808c3443cc49c670d0fb9bd64bd5d1ecd2b39b59898cb3fc4"} Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.431824 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6t92p"] Dec 04 14:18:01 crc kubenswrapper[4715]: E1204 14:18:01.432228 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="868a1540-37ed-4f8c-8a0b-2d44f99ab78b" containerName="ovn-config" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.432247 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="868a1540-37ed-4f8c-8a0b-2d44f99ab78b" containerName="ovn-config" Dec 04 14:18:01 crc kubenswrapper[4715]: E1204 14:18:01.432264 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2386cbfc-2eba-4ace-a56a-0e203b7b9229" containerName="swift-ring-rebalance" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.432270 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2386cbfc-2eba-4ace-a56a-0e203b7b9229" containerName="swift-ring-rebalance" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.432456 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="868a1540-37ed-4f8c-8a0b-2d44f99ab78b" containerName="ovn-config" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.432482 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2386cbfc-2eba-4ace-a56a-0e203b7b9229" containerName="swift-ring-rebalance" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.433157 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.462001 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6t92p"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.519677 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-p2wq9"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.520940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.540060 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-p2wq9"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.551248 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b8f3-account-create-update-xllmc"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.552417 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.561020 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.568717 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8496bf97-de9b-4438-a2d6-2af2046ebdd9-operator-scripts\") pod \"cinder-db-create-6t92p\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.568857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt5nt\" (UniqueName: \"kubernetes.io/projected/8496bf97-de9b-4438-a2d6-2af2046ebdd9-kube-api-access-rt5nt\") pod \"cinder-db-create-6t92p\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.593621 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b8f3-account-create-update-xllmc"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.652320 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c614-account-create-update-52t56"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.654272 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.670615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8496bf97-de9b-4438-a2d6-2af2046ebdd9-operator-scripts\") pod \"cinder-db-create-6t92p\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.670681 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2lrl\" (UniqueName: \"kubernetes.io/projected/ee23d5e5-7936-44fe-8d9a-041a52063147-kube-api-access-r2lrl\") pod \"barbican-b8f3-account-create-update-xllmc\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.670738 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee23d5e5-7936-44fe-8d9a-041a52063147-operator-scripts\") pod \"barbican-b8f3-account-create-update-xllmc\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.670777 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt5nt\" (UniqueName: \"kubernetes.io/projected/8496bf97-de9b-4438-a2d6-2af2046ebdd9-kube-api-access-rt5nt\") pod \"cinder-db-create-6t92p\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.670815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5607a052-6021-469b-a120-b95678e39c43-operator-scripts\") pod \"barbican-db-create-p2wq9\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.670895 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqvd6\" (UniqueName: \"kubernetes.io/projected/5607a052-6021-469b-a120-b95678e39c43-kube-api-access-dqvd6\") pod \"barbican-db-create-p2wq9\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.671885 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c614-account-create-update-52t56"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.672103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8496bf97-de9b-4438-a2d6-2af2046ebdd9-operator-scripts\") pod \"cinder-db-create-6t92p\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.674417 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.710149 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt5nt\" (UniqueName: \"kubernetes.io/projected/8496bf97-de9b-4438-a2d6-2af2046ebdd9-kube-api-access-rt5nt\") pod \"cinder-db-create-6t92p\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.738686 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gh6ns"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.745018 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.748471 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.748713 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.748911 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.749062 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bskrn" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.749958 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gh6ns"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.751168 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.772587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5607a052-6021-469b-a120-b95678e39c43-operator-scripts\") pod \"barbican-db-create-p2wq9\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.772711 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqvd6\" (UniqueName: \"kubernetes.io/projected/5607a052-6021-469b-a120-b95678e39c43-kube-api-access-dqvd6\") pod \"barbican-db-create-p2wq9\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.772748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-operator-scripts\") pod \"cinder-c614-account-create-update-52t56\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.772779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn2fs\" (UniqueName: \"kubernetes.io/projected/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-kube-api-access-qn2fs\") pod \"cinder-c614-account-create-update-52t56\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.772811 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2lrl\" (UniqueName: \"kubernetes.io/projected/ee23d5e5-7936-44fe-8d9a-041a52063147-kube-api-access-r2lrl\") pod \"barbican-b8f3-account-create-update-xllmc\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.772856 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee23d5e5-7936-44fe-8d9a-041a52063147-operator-scripts\") pod \"barbican-b8f3-account-create-update-xllmc\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.773813 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee23d5e5-7936-44fe-8d9a-041a52063147-operator-scripts\") pod \"barbican-b8f3-account-create-update-xllmc\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.773884 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5607a052-6021-469b-a120-b95678e39c43-operator-scripts\") pod \"barbican-db-create-p2wq9\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.816543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2lrl\" (UniqueName: \"kubernetes.io/projected/ee23d5e5-7936-44fe-8d9a-041a52063147-kube-api-access-r2lrl\") pod \"barbican-b8f3-account-create-update-xllmc\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.819082 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqvd6\" (UniqueName: \"kubernetes.io/projected/5607a052-6021-469b-a120-b95678e39c43-kube-api-access-dqvd6\") pod \"barbican-db-create-p2wq9\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.823544 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-8j6np"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.825094 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.844128 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.855045 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8j6np"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.873197 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.873614 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-config-data\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.873751 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jgkz\" (UniqueName: \"kubernetes.io/projected/ffdd4a77-1d9f-4f90-b604-9182c576913f-kube-api-access-7jgkz\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.873864 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-operator-scripts\") pod \"cinder-c614-account-create-update-52t56\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.873927 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn2fs\" (UniqueName: \"kubernetes.io/projected/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-kube-api-access-qn2fs\") pod \"cinder-c614-account-create-update-52t56\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.873948 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-combined-ca-bundle\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.875201 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-operator-scripts\") pod \"cinder-c614-account-create-update-52t56\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.895529 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn2fs\" (UniqueName: \"kubernetes.io/projected/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-kube-api-access-qn2fs\") pod \"cinder-c614-account-create-update-52t56\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.946401 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ec3a-account-create-update-p9mp4"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.949759 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.952709 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.968596 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ec3a-account-create-update-p9mp4"] Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.976250 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph7cq\" (UniqueName: \"kubernetes.io/projected/9f42c793-540b-4f39-9325-54858d52bc58-kube-api-access-ph7cq\") pod \"neutron-db-create-8j6np\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.976303 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-config-data\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.976335 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jgkz\" (UniqueName: \"kubernetes.io/projected/ffdd4a77-1d9f-4f90-b604-9182c576913f-kube-api-access-7jgkz\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.976352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f42c793-540b-4f39-9325-54858d52bc58-operator-scripts\") pod \"neutron-db-create-8j6np\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.976704 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-combined-ca-bundle\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.984019 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-config-data\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.984754 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.993157 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jgkz\" (UniqueName: \"kubernetes.io/projected/ffdd4a77-1d9f-4f90-b604-9182c576913f-kube-api-access-7jgkz\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:01 crc kubenswrapper[4715]: I1204 14:18:01.994776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-combined-ca-bundle\") pod \"keystone-db-sync-gh6ns\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.079106 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f42c793-540b-4f39-9325-54858d52bc58-operator-scripts\") pod \"neutron-db-create-8j6np\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.079186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv79d\" (UniqueName: \"kubernetes.io/projected/439588bc-077f-40ac-9b7c-684b1ad07055-kube-api-access-pv79d\") pod \"neutron-ec3a-account-create-update-p9mp4\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.079232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439588bc-077f-40ac-9b7c-684b1ad07055-operator-scripts\") pod \"neutron-ec3a-account-create-update-p9mp4\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.079346 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph7cq\" (UniqueName: \"kubernetes.io/projected/9f42c793-540b-4f39-9325-54858d52bc58-kube-api-access-ph7cq\") pod \"neutron-db-create-8j6np\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.080381 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f42c793-540b-4f39-9325-54858d52bc58-operator-scripts\") pod \"neutron-db-create-8j6np\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.100623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph7cq\" (UniqueName: \"kubernetes.io/projected/9f42c793-540b-4f39-9325-54858d52bc58-kube-api-access-ph7cq\") pod \"neutron-db-create-8j6np\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.161842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.183214 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv79d\" (UniqueName: \"kubernetes.io/projected/439588bc-077f-40ac-9b7c-684b1ad07055-kube-api-access-pv79d\") pod \"neutron-ec3a-account-create-update-p9mp4\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.183264 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439588bc-077f-40ac-9b7c-684b1ad07055-operator-scripts\") pod \"neutron-ec3a-account-create-update-p9mp4\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.183907 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439588bc-077f-40ac-9b7c-684b1ad07055-operator-scripts\") pod \"neutron-ec3a-account-create-update-p9mp4\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.184712 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.209122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv79d\" (UniqueName: \"kubernetes.io/projected/439588bc-077f-40ac-9b7c-684b1ad07055-kube-api-access-pv79d\") pod \"neutron-ec3a-account-create-update-p9mp4\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.350956 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.376867 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6t92p"] Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.442795 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c614-account-create-update-52t56"] Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.516618 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-p2wq9"] Dec 04 14:18:02 crc kubenswrapper[4715]: I1204 14:18:02.524305 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b8f3-account-create-update-xllmc"] Dec 04 14:18:02 crc kubenswrapper[4715]: W1204 14:18:02.655972 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8496bf97_de9b_4438_a2d6_2af2046ebdd9.slice/crio-9d48526d34bd9d62cc02a7d4f7ab6ba2ac111e266d5b70b4743c91d92a3b3cce WatchSource:0}: Error finding container 9d48526d34bd9d62cc02a7d4f7ab6ba2ac111e266d5b70b4743c91d92a3b3cce: Status 404 returned error can't find the container with id 9d48526d34bd9d62cc02a7d4f7ab6ba2ac111e266d5b70b4743c91d92a3b3cce Dec 04 14:18:02 crc kubenswrapper[4715]: W1204 14:18:02.667755 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5607a052_6021_469b_a120_b95678e39c43.slice/crio-ad76429c57362f0366245ad1eb6892529268560160d41bde26f369cbf1e296ed WatchSource:0}: Error finding container ad76429c57362f0366245ad1eb6892529268560160d41bde26f369cbf1e296ed: Status 404 returned error can't find the container with id ad76429c57362f0366245ad1eb6892529268560160d41bde26f369cbf1e296ed Dec 04 14:18:02 crc kubenswrapper[4715]: W1204 14:18:02.669263 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee23d5e5_7936_44fe_8d9a_041a52063147.slice/crio-6f63b16e396c321d71f667dc109f0d4f5b3ee34a73abbd3d904f9ea2ae665148 WatchSource:0}: Error finding container 6f63b16e396c321d71f667dc109f0d4f5b3ee34a73abbd3d904f9ea2ae665148: Status 404 returned error can't find the container with id 6f63b16e396c321d71f667dc109f0d4f5b3ee34a73abbd3d904f9ea2ae665148 Dec 04 14:18:02 crc kubenswrapper[4715]: W1204 14:18:02.672577 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a58c31a_8d3f_4425_bfdf_17c7c5f149aa.slice/crio-e45f402be8a9dc1f4f6dd8d59cbbdc3d83d8ba8ef773b97f01985d2c121acdbe WatchSource:0}: Error finding container e45f402be8a9dc1f4f6dd8d59cbbdc3d83d8ba8ef773b97f01985d2c121acdbe: Status 404 returned error can't find the container with id e45f402be8a9dc1f4f6dd8d59cbbdc3d83d8ba8ef773b97f01985d2c121acdbe Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.220782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6t92p" event={"ID":"8496bf97-de9b-4438-a2d6-2af2046ebdd9","Type":"ContainerStarted","Data":"9d48526d34bd9d62cc02a7d4f7ab6ba2ac111e266d5b70b4743c91d92a3b3cce"} Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.235214 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p2wq9" event={"ID":"5607a052-6021-469b-a120-b95678e39c43","Type":"ContainerStarted","Data":"f0de8923127566d140d82d25dcedf2bcc2a862e2abc3a3863d90fe6f950bf8e3"} Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.235269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p2wq9" event={"ID":"5607a052-6021-469b-a120-b95678e39c43","Type":"ContainerStarted","Data":"ad76429c57362f0366245ad1eb6892529268560160d41bde26f369cbf1e296ed"} Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.239074 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b8f3-account-create-update-xllmc" event={"ID":"ee23d5e5-7936-44fe-8d9a-041a52063147","Type":"ContainerStarted","Data":"6f63b16e396c321d71f667dc109f0d4f5b3ee34a73abbd3d904f9ea2ae665148"} Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.248646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c614-account-create-update-52t56" event={"ID":"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa","Type":"ContainerStarted","Data":"e45f402be8a9dc1f4f6dd8d59cbbdc3d83d8ba8ef773b97f01985d2c121acdbe"} Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.294768 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8j6np"] Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.422254 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gh6ns"] Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.514265 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ec3a-account-create-update-p9mp4"] Dec 04 14:18:03 crc kubenswrapper[4715]: I1204 14:18:03.522212 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-p2wq9" podStartSLOduration=2.522182217 podStartE2EDuration="2.522182217s" podCreationTimestamp="2025-12-04 14:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:03.392843674 +0000 UTC m=+1240.461561899" watchObservedRunningTime="2025-12-04 14:18:03.522182217 +0000 UTC m=+1240.590900432" Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.260366 4715 generic.go:334] "Generic (PLEG): container finished" podID="5607a052-6021-469b-a120-b95678e39c43" containerID="f0de8923127566d140d82d25dcedf2bcc2a862e2abc3a3863d90fe6f950bf8e3" exitCode=0 Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.260495 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p2wq9" event={"ID":"5607a052-6021-469b-a120-b95678e39c43","Type":"ContainerDied","Data":"f0de8923127566d140d82d25dcedf2bcc2a862e2abc3a3863d90fe6f950bf8e3"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.267473 4715 generic.go:334] "Generic (PLEG): container finished" podID="ee23d5e5-7936-44fe-8d9a-041a52063147" containerID="8b6007f3e8883b3289191a9f604bff9e99321a68097772512a143d00483a3292" exitCode=0 Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.267573 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b8f3-account-create-update-xllmc" event={"ID":"ee23d5e5-7936-44fe-8d9a-041a52063147","Type":"ContainerDied","Data":"8b6007f3e8883b3289191a9f604bff9e99321a68097772512a143d00483a3292"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.269811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gh6ns" event={"ID":"ffdd4a77-1d9f-4f90-b604-9182c576913f","Type":"ContainerStarted","Data":"bf8e5415ee8ebd1611ca312d5bbcb6bdaeef312196ac62b3efdcd4fe1aa46851"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.271456 4715 generic.go:334] "Generic (PLEG): container finished" podID="8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" containerID="4155d5f363c8bf805ab45145ec24681cecd4abe1a4dd874b59257211d3d28882" exitCode=0 Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.271600 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c614-account-create-update-52t56" event={"ID":"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa","Type":"ContainerDied","Data":"4155d5f363c8bf805ab45145ec24681cecd4abe1a4dd874b59257211d3d28882"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.273962 4715 generic.go:334] "Generic (PLEG): container finished" podID="8496bf97-de9b-4438-a2d6-2af2046ebdd9" containerID="da6c07fa2bd0fbab936b95f437fa6b0b4046b710cddbd71a18cbbcd6caa32ec3" exitCode=0 Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.274050 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6t92p" event={"ID":"8496bf97-de9b-4438-a2d6-2af2046ebdd9","Type":"ContainerDied","Data":"da6c07fa2bd0fbab936b95f437fa6b0b4046b710cddbd71a18cbbcd6caa32ec3"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.276640 4715 generic.go:334] "Generic (PLEG): container finished" podID="439588bc-077f-40ac-9b7c-684b1ad07055" containerID="a90919ab0e8b4d36775da8dc231de863c7940e1fddf6a502eba1483849e70fc9" exitCode=0 Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.276722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec3a-account-create-update-p9mp4" event={"ID":"439588bc-077f-40ac-9b7c-684b1ad07055","Type":"ContainerDied","Data":"a90919ab0e8b4d36775da8dc231de863c7940e1fddf6a502eba1483849e70fc9"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.276754 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec3a-account-create-update-p9mp4" event={"ID":"439588bc-077f-40ac-9b7c-684b1ad07055","Type":"ContainerStarted","Data":"3e648987af8bbc510501eec5e3104174285c0199ca9026ec21274c7e52525935"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.281163 4715 generic.go:334] "Generic (PLEG): container finished" podID="9f42c793-540b-4f39-9325-54858d52bc58" containerID="517cc83fdf050232a010ef797a10434f7aecacea989c3043896f3ee79e28f824" exitCode=0 Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.281274 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8j6np" event={"ID":"9f42c793-540b-4f39-9325-54858d52bc58","Type":"ContainerDied","Data":"517cc83fdf050232a010ef797a10434f7aecacea989c3043896f3ee79e28f824"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.281309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8j6np" event={"ID":"9f42c793-540b-4f39-9325-54858d52bc58","Type":"ContainerStarted","Data":"7ea81b18564d51a0fef5e3ebe3bcc743d94fc14cf912b2ff8c9659c8971ee0f3"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.289295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"113f79a89c02930e6fd386d45544ef39e76e9116a0d6dc07d77feca9d0ad61ed"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.289551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"e7da2f0640f9805d6ab1e352367e5fae35aa02e678150891b5ab9899d0ac2119"} Dec 04 14:18:04 crc kubenswrapper[4715]: I1204 14:18:04.289644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"cadde25fe632eddfee44fa8bfaf54c93c45c12f1297825a857f3c0ebe2590008"} Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.306067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"a4f6a22a2bc1819059b7abc60f6ca2ce79924c60eb53b9317e325c88d993dfdd"} Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.811499 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.968244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5607a052-6021-469b-a120-b95678e39c43-operator-scripts\") pod \"5607a052-6021-469b-a120-b95678e39c43\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.968458 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqvd6\" (UniqueName: \"kubernetes.io/projected/5607a052-6021-469b-a120-b95678e39c43-kube-api-access-dqvd6\") pod \"5607a052-6021-469b-a120-b95678e39c43\" (UID: \"5607a052-6021-469b-a120-b95678e39c43\") " Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.969371 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5607a052-6021-469b-a120-b95678e39c43-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5607a052-6021-469b-a120-b95678e39c43" (UID: "5607a052-6021-469b-a120-b95678e39c43"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.969900 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5607a052-6021-469b-a120-b95678e39c43-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:05 crc kubenswrapper[4715]: I1204 14:18:05.974861 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5607a052-6021-469b-a120-b95678e39c43-kube-api-access-dqvd6" (OuterVolumeSpecName: "kube-api-access-dqvd6") pod "5607a052-6021-469b-a120-b95678e39c43" (UID: "5607a052-6021-469b-a120-b95678e39c43"). InnerVolumeSpecName "kube-api-access-dqvd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.071909 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqvd6\" (UniqueName: \"kubernetes.io/projected/5607a052-6021-469b-a120-b95678e39c43-kube-api-access-dqvd6\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.313984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p2wq9" event={"ID":"5607a052-6021-469b-a120-b95678e39c43","Type":"ContainerDied","Data":"ad76429c57362f0366245ad1eb6892529268560160d41bde26f369cbf1e296ed"} Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.314404 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad76429c57362f0366245ad1eb6892529268560160d41bde26f369cbf1e296ed" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.314153 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p2wq9" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.563664 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.573127 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.592593 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.606072 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.680523 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2lrl\" (UniqueName: \"kubernetes.io/projected/ee23d5e5-7936-44fe-8d9a-041a52063147-kube-api-access-r2lrl\") pod \"ee23d5e5-7936-44fe-8d9a-041a52063147\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.680657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee23d5e5-7936-44fe-8d9a-041a52063147-operator-scripts\") pod \"ee23d5e5-7936-44fe-8d9a-041a52063147\" (UID: \"ee23d5e5-7936-44fe-8d9a-041a52063147\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.680759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439588bc-077f-40ac-9b7c-684b1ad07055-operator-scripts\") pod \"439588bc-077f-40ac-9b7c-684b1ad07055\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.680806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv79d\" (UniqueName: \"kubernetes.io/projected/439588bc-077f-40ac-9b7c-684b1ad07055-kube-api-access-pv79d\") pod \"439588bc-077f-40ac-9b7c-684b1ad07055\" (UID: \"439588bc-077f-40ac-9b7c-684b1ad07055\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.680851 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8496bf97-de9b-4438-a2d6-2af2046ebdd9-operator-scripts\") pod \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.680905 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt5nt\" (UniqueName: \"kubernetes.io/projected/8496bf97-de9b-4438-a2d6-2af2046ebdd9-kube-api-access-rt5nt\") pod \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\" (UID: \"8496bf97-de9b-4438-a2d6-2af2046ebdd9\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.681502 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/439588bc-077f-40ac-9b7c-684b1ad07055-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "439588bc-077f-40ac-9b7c-684b1ad07055" (UID: "439588bc-077f-40ac-9b7c-684b1ad07055"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.681497 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee23d5e5-7936-44fe-8d9a-041a52063147-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee23d5e5-7936-44fe-8d9a-041a52063147" (UID: "ee23d5e5-7936-44fe-8d9a-041a52063147"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.681559 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8496bf97-de9b-4438-a2d6-2af2046ebdd9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8496bf97-de9b-4438-a2d6-2af2046ebdd9" (UID: "8496bf97-de9b-4438-a2d6-2af2046ebdd9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.685726 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8496bf97-de9b-4438-a2d6-2af2046ebdd9-kube-api-access-rt5nt" (OuterVolumeSpecName: "kube-api-access-rt5nt") pod "8496bf97-de9b-4438-a2d6-2af2046ebdd9" (UID: "8496bf97-de9b-4438-a2d6-2af2046ebdd9"). InnerVolumeSpecName "kube-api-access-rt5nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.686135 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/439588bc-077f-40ac-9b7c-684b1ad07055-kube-api-access-pv79d" (OuterVolumeSpecName: "kube-api-access-pv79d") pod "439588bc-077f-40ac-9b7c-684b1ad07055" (UID: "439588bc-077f-40ac-9b7c-684b1ad07055"). InnerVolumeSpecName "kube-api-access-pv79d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.686591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee23d5e5-7936-44fe-8d9a-041a52063147-kube-api-access-r2lrl" (OuterVolumeSpecName: "kube-api-access-r2lrl") pod "ee23d5e5-7936-44fe-8d9a-041a52063147" (UID: "ee23d5e5-7936-44fe-8d9a-041a52063147"). InnerVolumeSpecName "kube-api-access-r2lrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.782424 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph7cq\" (UniqueName: \"kubernetes.io/projected/9f42c793-540b-4f39-9325-54858d52bc58-kube-api-access-ph7cq\") pod \"9f42c793-540b-4f39-9325-54858d52bc58\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.782611 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f42c793-540b-4f39-9325-54858d52bc58-operator-scripts\") pod \"9f42c793-540b-4f39-9325-54858d52bc58\" (UID: \"9f42c793-540b-4f39-9325-54858d52bc58\") " Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783220 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439588bc-077f-40ac-9b7c-684b1ad07055-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783244 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv79d\" (UniqueName: \"kubernetes.io/projected/439588bc-077f-40ac-9b7c-684b1ad07055-kube-api-access-pv79d\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783262 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8496bf97-de9b-4438-a2d6-2af2046ebdd9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783275 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt5nt\" (UniqueName: \"kubernetes.io/projected/8496bf97-de9b-4438-a2d6-2af2046ebdd9-kube-api-access-rt5nt\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783287 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2lrl\" (UniqueName: \"kubernetes.io/projected/ee23d5e5-7936-44fe-8d9a-041a52063147-kube-api-access-r2lrl\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783299 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee23d5e5-7936-44fe-8d9a-041a52063147-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.783459 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f42c793-540b-4f39-9325-54858d52bc58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f42c793-540b-4f39-9325-54858d52bc58" (UID: "9f42c793-540b-4f39-9325-54858d52bc58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.790117 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f42c793-540b-4f39-9325-54858d52bc58-kube-api-access-ph7cq" (OuterVolumeSpecName: "kube-api-access-ph7cq") pod "9f42c793-540b-4f39-9325-54858d52bc58" (UID: "9f42c793-540b-4f39-9325-54858d52bc58"). InnerVolumeSpecName "kube-api-access-ph7cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.885019 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph7cq\" (UniqueName: \"kubernetes.io/projected/9f42c793-540b-4f39-9325-54858d52bc58-kube-api-access-ph7cq\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:06 crc kubenswrapper[4715]: I1204 14:18:06.885235 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f42c793-540b-4f39-9325-54858d52bc58-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.323830 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec3a-account-create-update-p9mp4" event={"ID":"439588bc-077f-40ac-9b7c-684b1ad07055","Type":"ContainerDied","Data":"3e648987af8bbc510501eec5e3104174285c0199ca9026ec21274c7e52525935"} Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.323871 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e648987af8bbc510501eec5e3104174285c0199ca9026ec21274c7e52525935" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.323873 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec3a-account-create-update-p9mp4" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.325738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8j6np" event={"ID":"9f42c793-540b-4f39-9325-54858d52bc58","Type":"ContainerDied","Data":"7ea81b18564d51a0fef5e3ebe3bcc743d94fc14cf912b2ff8c9659c8971ee0f3"} Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.325764 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ea81b18564d51a0fef5e3ebe3bcc743d94fc14cf912b2ff8c9659c8971ee0f3" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.325801 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8j6np" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.327132 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b8f3-account-create-update-xllmc" event={"ID":"ee23d5e5-7936-44fe-8d9a-041a52063147","Type":"ContainerDied","Data":"6f63b16e396c321d71f667dc109f0d4f5b3ee34a73abbd3d904f9ea2ae665148"} Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.327155 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b8f3-account-create-update-xllmc" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.327169 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f63b16e396c321d71f667dc109f0d4f5b3ee34a73abbd3d904f9ea2ae665148" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.330111 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6t92p" event={"ID":"8496bf97-de9b-4438-a2d6-2af2046ebdd9","Type":"ContainerDied","Data":"9d48526d34bd9d62cc02a7d4f7ab6ba2ac111e266d5b70b4743c91d92a3b3cce"} Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.330147 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d48526d34bd9d62cc02a7d4f7ab6ba2ac111e266d5b70b4743c91d92a3b3cce" Dec 04 14:18:07 crc kubenswrapper[4715]: I1204 14:18:07.330193 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6t92p" Dec 04 14:18:08 crc kubenswrapper[4715]: I1204 14:18:08.758248 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:18:08 crc kubenswrapper[4715]: I1204 14:18:08.758605 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.454779 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.639198 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn2fs\" (UniqueName: \"kubernetes.io/projected/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-kube-api-access-qn2fs\") pod \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.639447 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-operator-scripts\") pod \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\" (UID: \"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa\") " Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.640634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" (UID: "8a58c31a-8d3f-4425-bfdf-17c7c5f149aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.646084 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-kube-api-access-qn2fs" (OuterVolumeSpecName: "kube-api-access-qn2fs") pod "8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" (UID: "8a58c31a-8d3f-4425-bfdf-17c7c5f149aa"). InnerVolumeSpecName "kube-api-access-qn2fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.741270 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:09 crc kubenswrapper[4715]: I1204 14:18:09.741305 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn2fs\" (UniqueName: \"kubernetes.io/projected/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa-kube-api-access-qn2fs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.356524 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c614-account-create-update-52t56" Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.356523 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c614-account-create-update-52t56" event={"ID":"8a58c31a-8d3f-4425-bfdf-17c7c5f149aa","Type":"ContainerDied","Data":"e45f402be8a9dc1f4f6dd8d59cbbdc3d83d8ba8ef773b97f01985d2c121acdbe"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.356971 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e45f402be8a9dc1f4f6dd8d59cbbdc3d83d8ba8ef773b97f01985d2c121acdbe" Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.359629 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ddj77" event={"ID":"88b59e0b-f6fd-42e0-9780-7c13708c379c","Type":"ContainerStarted","Data":"d5ad74cc3099453913a147a19f61db4e1d3624950e0594fb15fb0f42e1711c03"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.364405 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"c8c91f46e86048d0483dd4efc0acbda8d20c54da91d60ca590302db409bb400f"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.364459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"07258903ba29aec39ae701b10d6ff7be62848dec4683551bddda03b732901602"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.364472 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"25774c114a2acd38f302d0f0768876716dff50721ed529110ea92174dfc69579"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.364483 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"634d830151878b676643d440dd7fb2cdcd37c5fbbcb235417848604563e5c098"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.366120 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gh6ns" event={"ID":"ffdd4a77-1d9f-4f90-b604-9182c576913f","Type":"ContainerStarted","Data":"4bfa2f10b9bddae5417d80c0b064c3bf49e30475d9b123ad0e559f8e6b3d6532"} Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.384814 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ddj77" podStartSLOduration=1.846938459 podStartE2EDuration="32.384793677s" podCreationTimestamp="2025-12-04 14:17:38 +0000 UTC" firstStartedPulling="2025-12-04 14:17:38.965528319 +0000 UTC m=+1216.034246534" lastFinishedPulling="2025-12-04 14:18:09.503383537 +0000 UTC m=+1246.572101752" observedRunningTime="2025-12-04 14:18:10.375951748 +0000 UTC m=+1247.444669983" watchObservedRunningTime="2025-12-04 14:18:10.384793677 +0000 UTC m=+1247.453511902" Dec 04 14:18:10 crc kubenswrapper[4715]: I1204 14:18:10.497504 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gh6ns" podStartSLOduration=3.506493223 podStartE2EDuration="9.497480529s" podCreationTimestamp="2025-12-04 14:18:01 +0000 UTC" firstStartedPulling="2025-12-04 14:18:03.513672686 +0000 UTC m=+1240.582390911" lastFinishedPulling="2025-12-04 14:18:09.504660002 +0000 UTC m=+1246.573378217" observedRunningTime="2025-12-04 14:18:10.406400462 +0000 UTC m=+1247.475118677" watchObservedRunningTime="2025-12-04 14:18:10.497480529 +0000 UTC m=+1247.566198744" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.428252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"612f29d950bb08d6054afabf120c1ed82de78bf9944cc6553e51f0e71a879a04"} Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.429296 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"27437d99af1aae78ae2d23ce2b14b7dd3ba837db3a6686ceb45ebfd472d2af61"} Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.429320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"548d91b6-2482-4624-843a-ea2898f7beb8","Type":"ContainerStarted","Data":"4c8ef9313deb8f524bfb9b78528a718cadaca8666a03ac9f740acf7ffc2eb0fb"} Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.474179 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=42.210769827 podStartE2EDuration="53.474147688s" podCreationTimestamp="2025-12-04 14:17:24 +0000 UTC" firstStartedPulling="2025-12-04 14:17:58.125817174 +0000 UTC m=+1235.194535389" lastFinishedPulling="2025-12-04 14:18:09.389195035 +0000 UTC m=+1246.457913250" observedRunningTime="2025-12-04 14:18:17.464868117 +0000 UTC m=+1254.533586332" watchObservedRunningTime="2025-12-04 14:18:17.474147688 +0000 UTC m=+1254.542865913" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.735731 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-jh6z2"] Dec 04 14:18:17 crc kubenswrapper[4715]: E1204 14:18:17.736171 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8496bf97-de9b-4438-a2d6-2af2046ebdd9" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736222 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8496bf97-de9b-4438-a2d6-2af2046ebdd9" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: E1204 14:18:17.736241 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f42c793-540b-4f39-9325-54858d52bc58" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736250 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f42c793-540b-4f39-9325-54858d52bc58" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: E1204 14:18:17.736266 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5607a052-6021-469b-a120-b95678e39c43" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736272 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5607a052-6021-469b-a120-b95678e39c43" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: E1204 14:18:17.736284 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee23d5e5-7936-44fe-8d9a-041a52063147" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736290 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee23d5e5-7936-44fe-8d9a-041a52063147" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: E1204 14:18:17.736301 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736307 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: E1204 14:18:17.736321 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439588bc-077f-40ac-9b7c-684b1ad07055" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736327 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="439588bc-077f-40ac-9b7c-684b1ad07055" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736523 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f42c793-540b-4f39-9325-54858d52bc58" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736560 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5607a052-6021-469b-a120-b95678e39c43" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736572 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="439588bc-077f-40ac-9b7c-684b1ad07055" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736581 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8496bf97-de9b-4438-a2d6-2af2046ebdd9" containerName="mariadb-database-create" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736596 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee23d5e5-7936-44fe-8d9a-041a52063147" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.736609 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" containerName="mariadb-account-create-update" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.737864 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.740510 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.761183 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-jh6z2"] Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.887048 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg267\" (UniqueName: \"kubernetes.io/projected/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-kube-api-access-vg267\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.887112 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-svc\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.887187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.887216 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.887279 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:17 crc kubenswrapper[4715]: I1204 14:18:17.887296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-config\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.989251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.989339 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-config\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.989417 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg267\" (UniqueName: \"kubernetes.io/projected/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-kube-api-access-vg267\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.989478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-svc\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.989586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.989644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.991191 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.992221 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.993027 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-svc\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:17.994011 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:18.061708 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-config\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:18.084627 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg267\" (UniqueName: \"kubernetes.io/projected/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-kube-api-access-vg267\") pod \"dnsmasq-dns-764c5664d7-jh6z2\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:18.356462 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:18 crc kubenswrapper[4715]: I1204 14:18:18.811698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-jh6z2"] Dec 04 14:18:19 crc kubenswrapper[4715]: I1204 14:18:19.480392 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerID="b3129c55556cd5b4248341ffc19e48bb5694a1fe6cae049e340ba38e6b8b2c28" exitCode=0 Dec 04 14:18:19 crc kubenswrapper[4715]: I1204 14:18:19.480608 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" event={"ID":"8b84114f-0fa6-4ce0-8a2b-435b30f043fa","Type":"ContainerDied","Data":"b3129c55556cd5b4248341ffc19e48bb5694a1fe6cae049e340ba38e6b8b2c28"} Dec 04 14:18:19 crc kubenswrapper[4715]: I1204 14:18:19.480636 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" event={"ID":"8b84114f-0fa6-4ce0-8a2b-435b30f043fa","Type":"ContainerStarted","Data":"24e0442817e2f041e7dbacd6ced4580e11c0a3d4af705c9faa3a039418d8ff12"} Dec 04 14:18:20 crc kubenswrapper[4715]: I1204 14:18:20.489143 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" event={"ID":"8b84114f-0fa6-4ce0-8a2b-435b30f043fa","Type":"ContainerStarted","Data":"73ef168c82ffff77100d52a565551c60602f684026edcfda34e103dbc82312e7"} Dec 04 14:18:20 crc kubenswrapper[4715]: I1204 14:18:20.489508 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:20 crc kubenswrapper[4715]: I1204 14:18:20.521947 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" podStartSLOduration=3.521928368 podStartE2EDuration="3.521928368s" podCreationTimestamp="2025-12-04 14:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:20.512692348 +0000 UTC m=+1257.581410563" watchObservedRunningTime="2025-12-04 14:18:20.521928368 +0000 UTC m=+1257.590646583" Dec 04 14:18:21 crc kubenswrapper[4715]: I1204 14:18:21.506761 4715 generic.go:334] "Generic (PLEG): container finished" podID="ffdd4a77-1d9f-4f90-b604-9182c576913f" containerID="4bfa2f10b9bddae5417d80c0b064c3bf49e30475d9b123ad0e559f8e6b3d6532" exitCode=0 Dec 04 14:18:21 crc kubenswrapper[4715]: I1204 14:18:21.506811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gh6ns" event={"ID":"ffdd4a77-1d9f-4f90-b604-9182c576913f","Type":"ContainerDied","Data":"4bfa2f10b9bddae5417d80c0b064c3bf49e30475d9b123ad0e559f8e6b3d6532"} Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.878611 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.900760 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-config-data\") pod \"ffdd4a77-1d9f-4f90-b604-9182c576913f\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.900853 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-combined-ca-bundle\") pod \"ffdd4a77-1d9f-4f90-b604-9182c576913f\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.901026 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jgkz\" (UniqueName: \"kubernetes.io/projected/ffdd4a77-1d9f-4f90-b604-9182c576913f-kube-api-access-7jgkz\") pod \"ffdd4a77-1d9f-4f90-b604-9182c576913f\" (UID: \"ffdd4a77-1d9f-4f90-b604-9182c576913f\") " Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.906908 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffdd4a77-1d9f-4f90-b604-9182c576913f-kube-api-access-7jgkz" (OuterVolumeSpecName: "kube-api-access-7jgkz") pod "ffdd4a77-1d9f-4f90-b604-9182c576913f" (UID: "ffdd4a77-1d9f-4f90-b604-9182c576913f"). InnerVolumeSpecName "kube-api-access-7jgkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.926156 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffdd4a77-1d9f-4f90-b604-9182c576913f" (UID: "ffdd4a77-1d9f-4f90-b604-9182c576913f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:22 crc kubenswrapper[4715]: I1204 14:18:22.947896 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-config-data" (OuterVolumeSpecName: "config-data") pod "ffdd4a77-1d9f-4f90-b604-9182c576913f" (UID: "ffdd4a77-1d9f-4f90-b604-9182c576913f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.003303 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jgkz\" (UniqueName: \"kubernetes.io/projected/ffdd4a77-1d9f-4f90-b604-9182c576913f-kube-api-access-7jgkz\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.003591 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.003657 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffdd4a77-1d9f-4f90-b604-9182c576913f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.528386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gh6ns" event={"ID":"ffdd4a77-1d9f-4f90-b604-9182c576913f","Type":"ContainerDied","Data":"bf8e5415ee8ebd1611ca312d5bbcb6bdaeef312196ac62b3efdcd4fe1aa46851"} Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.528434 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf8e5415ee8ebd1611ca312d5bbcb6bdaeef312196ac62b3efdcd4fe1aa46851" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.528500 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gh6ns" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.880794 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-jh6z2"] Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.881074 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerName="dnsmasq-dns" containerID="cri-o://73ef168c82ffff77100d52a565551c60602f684026edcfda34e103dbc82312e7" gracePeriod=10 Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.935010 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7zbvm"] Dec 04 14:18:23 crc kubenswrapper[4715]: E1204 14:18:23.936235 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffdd4a77-1d9f-4f90-b604-9182c576913f" containerName="keystone-db-sync" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.936257 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffdd4a77-1d9f-4f90-b604-9182c576913f" containerName="keystone-db-sync" Dec 04 14:18:23 crc kubenswrapper[4715]: I1204 14:18:23.970437 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffdd4a77-1d9f-4f90-b604-9182c576913f" containerName="keystone-db-sync" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.027687 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7zbvm"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.027799 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.092146 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-z2s4m"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.093350 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.097023 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.097233 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.097424 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bskrn" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.097535 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.097732 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.106214 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.106316 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxsxg\" (UniqueName: \"kubernetes.io/projected/11dc0691-4593-4258-9bb5-4239c11cf835-kube-api-access-sxsxg\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.106359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.106391 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-svc\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.106486 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-config\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.106523 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.131097 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-z2s4m"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.199234 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5978d7d969-7vwfh"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.200796 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.205376 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-dt96p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.205401 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.205596 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.205781 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-credential-keys\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207797 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-scripts\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207824 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-config\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-combined-ca-bundle\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207867 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207886 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207923 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-config-data\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxsxg\" (UniqueName: \"kubernetes.io/projected/11dc0691-4593-4258-9bb5-4239c11cf835-kube-api-access-sxsxg\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.207993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-svc\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.208022 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-fernet-keys\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.208093 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l25vj\" (UniqueName: \"kubernetes.io/projected/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-kube-api-access-l25vj\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.208980 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-config\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.209779 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.209892 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.210057 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-cgdwm"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.210390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-svc\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.210694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.211207 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.222663 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zvl22" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.222839 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.222985 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.225100 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5978d7d969-7vwfh"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.236319 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cgdwm"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.250838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxsxg\" (UniqueName: \"kubernetes.io/projected/11dc0691-4593-4258-9bb5-4239c11cf835-kube-api-access-sxsxg\") pod \"dnsmasq-dns-5959f8865f-7zbvm\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.304322 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.306662 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.308982 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309060 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a909008-f754-49fb-87cb-dc0d09c91bfd-logs\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309195 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-config-data\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-etc-machine-id\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-credential-keys\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-scripts\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309400 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-scripts\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-combined-ca-bundle\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309485 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-combined-ca-bundle\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309515 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-db-sync-config-data\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309557 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a909008-f754-49fb-87cb-dc0d09c91bfd-horizon-secret-key\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-config-data\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp9nm\" (UniqueName: \"kubernetes.io/projected/0a909008-f754-49fb-87cb-dc0d09c91bfd-kube-api-access-bp9nm\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jqbz\" (UniqueName: \"kubernetes.io/projected/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-kube-api-access-2jqbz\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-scripts\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309791 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-fernet-keys\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309833 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-config-data\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.309865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l25vj\" (UniqueName: \"kubernetes.io/projected/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-kube-api-access-l25vj\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.310045 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.317813 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-config-data\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.325092 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-fernet-keys\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.338908 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-scripts\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.340054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-credential-keys\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.346312 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l25vj\" (UniqueName: \"kubernetes.io/projected/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-kube-api-access-l25vj\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.354197 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-combined-ca-bundle\") pod \"keystone-bootstrap-z2s4m\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.354288 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.365472 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.367412 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-8bbsj"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.373958 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.379779 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-j8psz" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.379985 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a909008-f754-49fb-87cb-dc0d09c91bfd-logs\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412222 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-log-httpd\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-config-data\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412306 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-etc-machine-id\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412373 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-scripts\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412425 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412452 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-combined-ca-bundle\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqtrm\" (UniqueName: \"kubernetes.io/projected/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-kube-api-access-tqtrm\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412504 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-db-sync-config-data\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412547 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a909008-f754-49fb-87cb-dc0d09c91bfd-horizon-secret-key\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412605 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp9nm\" (UniqueName: \"kubernetes.io/projected/0a909008-f754-49fb-87cb-dc0d09c91bfd-kube-api-access-bp9nm\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412638 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jqbz\" (UniqueName: \"kubernetes.io/projected/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-kube-api-access-2jqbz\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412660 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-scripts\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412704 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-run-httpd\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412747 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-scripts\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412765 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-config-data\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.412784 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-config-data\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.413287 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a909008-f754-49fb-87cb-dc0d09c91bfd-logs\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.414231 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-etc-machine-id\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.415353 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-scripts\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.422321 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a909008-f754-49fb-87cb-dc0d09c91bfd-horizon-secret-key\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.425015 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.428946 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-db-sync-config-data\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.433501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-scripts\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.434530 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-config-data\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.434829 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-config-data\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.458783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-combined-ca-bundle\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.466093 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp9nm\" (UniqueName: \"kubernetes.io/projected/0a909008-f754-49fb-87cb-dc0d09c91bfd-kube-api-access-bp9nm\") pod \"horizon-5978d7d969-7vwfh\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569675 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-db-sync-config-data\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569793 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569809 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8bbsj"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569816 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqtrm\" (UniqueName: \"kubernetes.io/projected/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-kube-api-access-tqtrm\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569916 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8g8g\" (UniqueName: \"kubernetes.io/projected/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-kube-api-access-g8g8g\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-run-httpd\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.569984 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-scripts\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.570006 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-config-data\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.575380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.578148 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-5mrnk"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.579367 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.580113 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-run-httpd\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.580338 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-combined-ca-bundle\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.580384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-log-httpd\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.580997 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.581739 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-log-httpd\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.583847 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.583946 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.584172 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ck2jj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.584314 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.600105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-config-data\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.613719 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jqbz\" (UniqueName: \"kubernetes.io/projected/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-kube-api-access-2jqbz\") pod \"cinder-db-sync-cgdwm\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.633151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-scripts\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.639302 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5mrnk"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.659566 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerID="73ef168c82ffff77100d52a565551c60602f684026edcfda34e103dbc82312e7" exitCode=0 Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.659624 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" event={"ID":"8b84114f-0fa6-4ce0-8a2b-435b30f043fa","Type":"ContainerDied","Data":"73ef168c82ffff77100d52a565551c60602f684026edcfda34e103dbc82312e7"} Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.665634 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqtrm\" (UniqueName: \"kubernetes.io/projected/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-kube-api-access-tqtrm\") pod \"ceilometer-0\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.677385 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8599dd67c5-qs56p"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.678974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.684734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-logs\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.684794 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-config\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.684825 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8g8g\" (UniqueName: \"kubernetes.io/projected/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-kube-api-access-g8g8g\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.684881 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-config-data\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.684961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-scripts\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.685012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-combined-ca-bundle\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.685073 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-combined-ca-bundle\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.685104 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxdvg\" (UniqueName: \"kubernetes.io/projected/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-kube-api-access-fxdvg\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.685132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-db-sync-config-data\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.685174 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-horizon-secret-key\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.685203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2psbw\" (UniqueName: \"kubernetes.io/projected/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-kube-api-access-2psbw\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.697501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-combined-ca-bundle\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.705541 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-db-sync-config-data\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.729219 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8g8g\" (UniqueName: \"kubernetes.io/projected/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-kube-api-access-g8g8g\") pod \"barbican-db-sync-8bbsj\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.756124 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8599dd67c5-qs56p"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.776898 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7zbvm"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.785961 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-s5j8r"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788162 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-logs\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788215 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-config\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788264 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-config-data\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788293 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-scripts\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788339 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-combined-ca-bundle\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxdvg\" (UniqueName: \"kubernetes.io/projected/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-kube-api-access-fxdvg\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788398 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-horizon-secret-key\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788418 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2psbw\" (UniqueName: \"kubernetes.io/projected/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-kube-api-access-2psbw\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.788951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-logs\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.795902 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-scripts\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.796012 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-config-data\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.799008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-horizon-secret-key\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.800681 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s5j8r"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.800838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.800918 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-config\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.804143 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.804502 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.804707 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qljnv" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.830265 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-rjgzw"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.830610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-combined-ca-bundle\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.832076 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.832118 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2psbw\" (UniqueName: \"kubernetes.io/projected/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-kube-api-access-2psbw\") pod \"horizon-8599dd67c5-qs56p\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.832883 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxdvg\" (UniqueName: \"kubernetes.io/projected/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-kube-api-access-fxdvg\") pod \"neutron-db-sync-5mrnk\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.909744 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.909893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.909961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs8gn\" (UniqueName: \"kubernetes.io/projected/5cc0cd44-94c5-41c6-a11b-06faa5c78299-kube-api-access-cs8gn\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh7fj\" (UniqueName: \"kubernetes.io/projected/7e311957-d00b-43da-a9dc-ca48386dfa88-kube-api-access-nh7fj\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910103 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-combined-ca-bundle\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910120 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-scripts\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910202 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910357 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e311957-d00b-43da-a9dc-ca48386dfa88-logs\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910451 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-config\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910489 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.910634 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-config-data\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.922017 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-rjgzw"] Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.939020 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.962425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:18:24 crc kubenswrapper[4715]: I1204 14:18:24.985868 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.020412 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-svc\") pod \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.020828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-config\") pod \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.020866 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg267\" (UniqueName: \"kubernetes.io/projected/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-kube-api-access-vg267\") pod \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.020938 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-swift-storage-0\") pod \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.020985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-sb\") pod \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-nb\") pod \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\" (UID: \"8b84114f-0fa6-4ce0-8a2b-435b30f043fa\") " Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-combined-ca-bundle\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021414 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-scripts\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e311957-d00b-43da-a9dc-ca48386dfa88-logs\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021561 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-config\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-config-data\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021617 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021678 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021730 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs8gn\" (UniqueName: \"kubernetes.io/projected/5cc0cd44-94c5-41c6-a11b-06faa5c78299-kube-api-access-cs8gn\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.021794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh7fj\" (UniqueName: \"kubernetes.io/projected/7e311957-d00b-43da-a9dc-ca48386dfa88-kube-api-access-nh7fj\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.023866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.029862 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.030251 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e311957-d00b-43da-a9dc-ca48386dfa88-logs\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.030849 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-config\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.031493 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.035330 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-kube-api-access-vg267" (OuterVolumeSpecName: "kube-api-access-vg267") pod "8b84114f-0fa6-4ce0-8a2b-435b30f043fa" (UID: "8b84114f-0fa6-4ce0-8a2b-435b30f043fa"). InnerVolumeSpecName "kube-api-access-vg267". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.042356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.043458 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-config-data\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.045284 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-combined-ca-bundle\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.048676 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.050587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-scripts\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.074695 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh7fj\" (UniqueName: \"kubernetes.io/projected/7e311957-d00b-43da-a9dc-ca48386dfa88-kube-api-access-nh7fj\") pod \"placement-db-sync-s5j8r\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.082661 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs8gn\" (UniqueName: \"kubernetes.io/projected/5cc0cd44-94c5-41c6-a11b-06faa5c78299-kube-api-access-cs8gn\") pod \"dnsmasq-dns-58dd9ff6bc-rjgzw\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.123425 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg267\" (UniqueName: \"kubernetes.io/projected/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-kube-api-access-vg267\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.149991 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s5j8r" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.164863 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.166792 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-config" (OuterVolumeSpecName: "config") pod "8b84114f-0fa6-4ce0-8a2b-435b30f043fa" (UID: "8b84114f-0fa6-4ce0-8a2b-435b30f043fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.201397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8b84114f-0fa6-4ce0-8a2b-435b30f043fa" (UID: "8b84114f-0fa6-4ce0-8a2b-435b30f043fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.229289 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.229339 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.235106 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8b84114f-0fa6-4ce0-8a2b-435b30f043fa" (UID: "8b84114f-0fa6-4ce0-8a2b-435b30f043fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.250357 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8b84114f-0fa6-4ce0-8a2b-435b30f043fa" (UID: "8b84114f-0fa6-4ce0-8a2b-435b30f043fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.315110 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8b84114f-0fa6-4ce0-8a2b-435b30f043fa" (UID: "8b84114f-0fa6-4ce0-8a2b-435b30f043fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.336061 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.336093 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.336108 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b84114f-0fa6-4ce0-8a2b-435b30f043fa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.359854 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7zbvm"] Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.427000 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-z2s4m"] Dec 04 14:18:25 crc kubenswrapper[4715]: W1204 14:18:25.445555 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d2c3b9d_ffd1_46db_a3f5_4f113daa2253.slice/crio-83c91bdca9cec1ebd4f70cbee8e003cb61e4c0f731b8c85ea9a557a256c5a6e2 WatchSource:0}: Error finding container 83c91bdca9cec1ebd4f70cbee8e003cb61e4c0f731b8c85ea9a557a256c5a6e2: Status 404 returned error can't find the container with id 83c91bdca9cec1ebd4f70cbee8e003cb61e4c0f731b8c85ea9a557a256c5a6e2 Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.545829 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5978d7d969-7vwfh"] Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.680792 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z2s4m" event={"ID":"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253","Type":"ContainerStarted","Data":"83c91bdca9cec1ebd4f70cbee8e003cb61e4c0f731b8c85ea9a557a256c5a6e2"} Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.684950 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" event={"ID":"8b84114f-0fa6-4ce0-8a2b-435b30f043fa","Type":"ContainerDied","Data":"24e0442817e2f041e7dbacd6ced4580e11c0a3d4af705c9faa3a039418d8ff12"} Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.685022 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-jh6z2" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.685054 4715 scope.go:117] "RemoveContainer" containerID="73ef168c82ffff77100d52a565551c60602f684026edcfda34e103dbc82312e7" Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.685947 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" event={"ID":"11dc0691-4593-4258-9bb5-4239c11cf835","Type":"ContainerStarted","Data":"b26c699e7f0561f997145964e4c846d50d03304c47110bb9250044e06db546eb"} Dec 04 14:18:25 crc kubenswrapper[4715]: I1204 14:18:25.689928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5978d7d969-7vwfh" event={"ID":"0a909008-f754-49fb-87cb-dc0d09c91bfd","Type":"ContainerStarted","Data":"6a6c7dbc250388b68fbbcc2528f953c90c29f88b244e2c257b35095b2da86c4d"} Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:25.741455 4715 scope.go:117] "RemoveContainer" containerID="b3129c55556cd5b4248341ffc19e48bb5694a1fe6cae049e340ba38e6b8b2c28" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:25.746937 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-jh6z2"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:25.758442 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-jh6z2"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:25.819762 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cgdwm"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:25.833802 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8bbsj"] Dec 04 14:18:27 crc kubenswrapper[4715]: W1204 14:18:25.836844 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ddb5944_20e2_4f19_96c4_ee93ac8cfa02.slice/crio-afe1eab2479b984b970b6b768e650c0d7366d5cff31df88c45304bd61310d3ba WatchSource:0}: Error finding container afe1eab2479b984b970b6b768e650c0d7366d5cff31df88c45304bd61310d3ba: Status 404 returned error can't find the container with id afe1eab2479b984b970b6b768e650c0d7366d5cff31df88c45304bd61310d3ba Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.700665 4715 generic.go:334] "Generic (PLEG): container finished" podID="11dc0691-4593-4258-9bb5-4239c11cf835" containerID="03ab2eba5cac372486d103fb818ea3ff047fbe58fd65f9ae6c413196d1bbc9fd" exitCode=0 Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.700770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" event={"ID":"11dc0691-4593-4258-9bb5-4239c11cf835","Type":"ContainerDied","Data":"03ab2eba5cac372486d103fb818ea3ff047fbe58fd65f9ae6c413196d1bbc9fd"} Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.703337 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8bbsj" event={"ID":"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02","Type":"ContainerStarted","Data":"afe1eab2479b984b970b6b768e650c0d7366d5cff31df88c45304bd61310d3ba"} Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.709551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z2s4m" event={"ID":"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253","Type":"ContainerStarted","Data":"d54ffff1c0ecf1f8e43275f596d60721c329f39ceab316eaa8dde2aa18974237"} Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.712400 4715 generic.go:334] "Generic (PLEG): container finished" podID="88b59e0b-f6fd-42e0-9780-7c13708c379c" containerID="d5ad74cc3099453913a147a19f61db4e1d3624950e0594fb15fb0f42e1711c03" exitCode=0 Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.712474 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ddj77" event={"ID":"88b59e0b-f6fd-42e0-9780-7c13708c379c","Type":"ContainerDied","Data":"d5ad74cc3099453913a147a19f61db4e1d3624950e0594fb15fb0f42e1711c03"} Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.714564 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cgdwm" event={"ID":"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0","Type":"ContainerStarted","Data":"47a3005461d3c23508fcdfc14b3c9c8582ca422ee1c6ba0242d407d0f374aa3b"} Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:26.758411 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-z2s4m" podStartSLOduration=3.758389931 podStartE2EDuration="3.758389931s" podCreationTimestamp="2025-12-04 14:18:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:26.740653641 +0000 UTC m=+1263.809371856" watchObservedRunningTime="2025-12-04 14:18:26.758389931 +0000 UTC m=+1263.827108146" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.202108 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" path="/var/lib/kubelet/pods/8b84114f-0fa6-4ce0-8a2b-435b30f043fa/volumes" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.413702 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5978d7d969-7vwfh"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.451206 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-57458fcb8f-w827q"] Dec 04 14:18:27 crc kubenswrapper[4715]: E1204 14:18:27.451552 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerName="init" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.451568 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerName="init" Dec 04 14:18:27 crc kubenswrapper[4715]: E1204 14:18:27.451624 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerName="dnsmasq-dns" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.451632 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerName="dnsmasq-dns" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.451818 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b84114f-0fa6-4ce0-8a2b-435b30f043fa" containerName="dnsmasq-dns" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.452831 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.474578 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57458fcb8f-w827q"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.570260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-scripts\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.570320 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm5x9\" (UniqueName: \"kubernetes.io/projected/a1c0859d-f9ae-457f-9c44-ea9891d7756b-kube-api-access-fm5x9\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.570359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-config-data\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.570393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1c0859d-f9ae-457f-9c44-ea9891d7756b-horizon-secret-key\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.570420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c0859d-f9ae-457f-9c44-ea9891d7756b-logs\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.593413 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.683266 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-config-data\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.683359 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1c0859d-f9ae-457f-9c44-ea9891d7756b-horizon-secret-key\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.683406 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c0859d-f9ae-457f-9c44-ea9891d7756b-logs\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.683536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-scripts\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.683590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm5x9\" (UniqueName: \"kubernetes.io/projected/a1c0859d-f9ae-457f-9c44-ea9891d7756b-kube-api-access-fm5x9\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.666630 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5mrnk"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.699741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1c0859d-f9ae-457f-9c44-ea9891d7756b-horizon-secret-key\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.810967 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c0859d-f9ae-457f-9c44-ea9891d7756b-logs\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.818915 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-scripts\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.819627 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-config-data\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.827195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm5x9\" (UniqueName: \"kubernetes.io/projected/a1c0859d-f9ae-457f-9c44-ea9891d7756b-kube-api-access-fm5x9\") pod \"horizon-57458fcb8f-w827q\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.875115 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8599dd67c5-qs56p"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.886647 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:18:27 crc kubenswrapper[4715]: I1204 14:18:27.905954 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s5j8r"] Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.021620 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-rjgzw"] Dec 04 14:18:28 crc kubenswrapper[4715]: W1204 14:18:28.097211 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b94f46e_2d75_479c_9ef6_b8fac3332e2f.slice/crio-2e7948acdab947fccfea2245893ee375640cadd491d806f2745ab1dd2937798c WatchSource:0}: Error finding container 2e7948acdab947fccfea2245893ee375640cadd491d806f2745ab1dd2937798c: Status 404 returned error can't find the container with id 2e7948acdab947fccfea2245893ee375640cadd491d806f2745ab1dd2937798c Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.104087 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.121909 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.227188 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-swift-storage-0\") pod \"11dc0691-4593-4258-9bb5-4239c11cf835\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.227482 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-svc\") pod \"11dc0691-4593-4258-9bb5-4239c11cf835\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.227586 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-config\") pod \"11dc0691-4593-4258-9bb5-4239c11cf835\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.227670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxsxg\" (UniqueName: \"kubernetes.io/projected/11dc0691-4593-4258-9bb5-4239c11cf835-kube-api-access-sxsxg\") pod \"11dc0691-4593-4258-9bb5-4239c11cf835\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.227749 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-sb\") pod \"11dc0691-4593-4258-9bb5-4239c11cf835\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.227856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-nb\") pod \"11dc0691-4593-4258-9bb5-4239c11cf835\" (UID: \"11dc0691-4593-4258-9bb5-4239c11cf835\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.250010 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11dc0691-4593-4258-9bb5-4239c11cf835-kube-api-access-sxsxg" (OuterVolumeSpecName: "kube-api-access-sxsxg") pod "11dc0691-4593-4258-9bb5-4239c11cf835" (UID: "11dc0691-4593-4258-9bb5-4239c11cf835"). InnerVolumeSpecName "kube-api-access-sxsxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.252959 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11dc0691-4593-4258-9bb5-4239c11cf835" (UID: "11dc0691-4593-4258-9bb5-4239c11cf835"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.266666 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11dc0691-4593-4258-9bb5-4239c11cf835" (UID: "11dc0691-4593-4258-9bb5-4239c11cf835"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.275307 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11dc0691-4593-4258-9bb5-4239c11cf835" (UID: "11dc0691-4593-4258-9bb5-4239c11cf835"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.301828 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-config" (OuterVolumeSpecName: "config") pod "11dc0691-4593-4258-9bb5-4239c11cf835" (UID: "11dc0691-4593-4258-9bb5-4239c11cf835"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.317384 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "11dc0691-4593-4258-9bb5-4239c11cf835" (UID: "11dc0691-4593-4258-9bb5-4239c11cf835"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.329149 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.329201 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.329213 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.329228 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.329240 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dc0691-4593-4258-9bb5-4239c11cf835-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.329251 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxsxg\" (UniqueName: \"kubernetes.io/projected/11dc0691-4593-4258-9bb5-4239c11cf835-kube-api-access-sxsxg\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.411193 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ddj77" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.581214 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c94g\" (UniqueName: \"kubernetes.io/projected/88b59e0b-f6fd-42e0-9780-7c13708c379c-kube-api-access-5c94g\") pod \"88b59e0b-f6fd-42e0-9780-7c13708c379c\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.581502 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-config-data\") pod \"88b59e0b-f6fd-42e0-9780-7c13708c379c\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.581542 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-combined-ca-bundle\") pod \"88b59e0b-f6fd-42e0-9780-7c13708c379c\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.581606 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-db-sync-config-data\") pod \"88b59e0b-f6fd-42e0-9780-7c13708c379c\" (UID: \"88b59e0b-f6fd-42e0-9780-7c13708c379c\") " Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.588154 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b59e0b-f6fd-42e0-9780-7c13708c379c-kube-api-access-5c94g" (OuterVolumeSpecName: "kube-api-access-5c94g") pod "88b59e0b-f6fd-42e0-9780-7c13708c379c" (UID: "88b59e0b-f6fd-42e0-9780-7c13708c379c"). InnerVolumeSpecName "kube-api-access-5c94g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.590396 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "88b59e0b-f6fd-42e0-9780-7c13708c379c" (UID: "88b59e0b-f6fd-42e0-9780-7c13708c379c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.616287 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88b59e0b-f6fd-42e0-9780-7c13708c379c" (UID: "88b59e0b-f6fd-42e0-9780-7c13708c379c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.643165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-config-data" (OuterVolumeSpecName: "config-data") pod "88b59e0b-f6fd-42e0-9780-7c13708c379c" (UID: "88b59e0b-f6fd-42e0-9780-7c13708c379c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.683807 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.683850 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c94g\" (UniqueName: \"kubernetes.io/projected/88b59e0b-f6fd-42e0-9780-7c13708c379c-kube-api-access-5c94g\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.683863 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.683871 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b59e0b-f6fd-42e0-9780-7c13708c379c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.916740 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" event={"ID":"5cc0cd44-94c5-41c6-a11b-06faa5c78299","Type":"ContainerStarted","Data":"ba31ffd36339fcaf717149af462289bd56a07d23a098e0e0e2515c629086adde"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.918923 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8599dd67c5-qs56p" event={"ID":"e8331eb7-653b-4a6c-ada8-f13dcf916a8a","Type":"ContainerStarted","Data":"ff111cd200947cb410cf03c26a74ef33d037230944e0df444cd3f0a71c55ca0c"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.926012 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5mrnk" event={"ID":"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1","Type":"ContainerStarted","Data":"db414aa09f958319c2529dfb137c1ccf39473c769761e83052843c41c224aec9"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.930944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerStarted","Data":"2e7948acdab947fccfea2245893ee375640cadd491d806f2745ab1dd2937798c"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.933380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ddj77" event={"ID":"88b59e0b-f6fd-42e0-9780-7c13708c379c","Type":"ContainerDied","Data":"280c6bb2a0e2cf2e2aac764992e2dc20ea4ccdc46ce3250eb5ccdc199b561984"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.933423 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="280c6bb2a0e2cf2e2aac764992e2dc20ea4ccdc46ce3250eb5ccdc199b561984" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.933473 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ddj77" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.937372 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s5j8r" event={"ID":"7e311957-d00b-43da-a9dc-ca48386dfa88","Type":"ContainerStarted","Data":"b2955330a8ff7a6416467c50f8c921b8d8ce53ee79c34adf4286d969d28a08ea"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.942225 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" event={"ID":"11dc0691-4593-4258-9bb5-4239c11cf835","Type":"ContainerDied","Data":"b26c699e7f0561f997145964e4c846d50d03304c47110bb9250044e06db546eb"} Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.942275 4715 scope.go:117] "RemoveContainer" containerID="03ab2eba5cac372486d103fb818ea3ff047fbe58fd65f9ae6c413196d1bbc9fd" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.942287 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7zbvm" Dec 04 14:18:28 crc kubenswrapper[4715]: I1204 14:18:28.959611 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57458fcb8f-w827q"] Dec 04 14:18:28 crc kubenswrapper[4715]: W1204 14:18:28.960934 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1c0859d_f9ae_457f_9c44_ea9891d7756b.slice/crio-f951019dbe81cf726cb7dc69baa250b984a5ccc0eae9f3b7aa00edabc6ea9306 WatchSource:0}: Error finding container f951019dbe81cf726cb7dc69baa250b984a5ccc0eae9f3b7aa00edabc6ea9306: Status 404 returned error can't find the container with id f951019dbe81cf726cb7dc69baa250b984a5ccc0eae9f3b7aa00edabc6ea9306 Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.038691 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7zbvm"] Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.045596 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7zbvm"] Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.204586 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11dc0691-4593-4258-9bb5-4239c11cf835" path="/var/lib/kubelet/pods/11dc0691-4593-4258-9bb5-4239c11cf835/volumes" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.369234 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-rjgzw"] Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.390589 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z4zqf"] Dec 04 14:18:29 crc kubenswrapper[4715]: E1204 14:18:29.391976 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11dc0691-4593-4258-9bb5-4239c11cf835" containerName="init" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.392061 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="11dc0691-4593-4258-9bb5-4239c11cf835" containerName="init" Dec 04 14:18:29 crc kubenswrapper[4715]: E1204 14:18:29.392121 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b59e0b-f6fd-42e0-9780-7c13708c379c" containerName="glance-db-sync" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.392193 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b59e0b-f6fd-42e0-9780-7c13708c379c" containerName="glance-db-sync" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.392428 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b59e0b-f6fd-42e0-9780-7c13708c379c" containerName="glance-db-sync" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.392515 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="11dc0691-4593-4258-9bb5-4239c11cf835" containerName="init" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.393543 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.414591 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z4zqf"] Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.516383 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.516473 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhtgt\" (UniqueName: \"kubernetes.io/projected/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-kube-api-access-lhtgt\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.516513 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.516561 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-config\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.516608 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.516671 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.619342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhtgt\" (UniqueName: \"kubernetes.io/projected/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-kube-api-access-lhtgt\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.619427 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.619479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-config\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.619514 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.619566 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.619629 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.620819 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.621056 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-config\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.621583 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.621788 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.622130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.654953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhtgt\" (UniqueName: \"kubernetes.io/projected/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-kube-api-access-lhtgt\") pod \"dnsmasq-dns-785d8bcb8c-z4zqf\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.727519 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.986185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57458fcb8f-w827q" event={"ID":"a1c0859d-f9ae-457f-9c44-ea9891d7756b","Type":"ContainerStarted","Data":"f951019dbe81cf726cb7dc69baa250b984a5ccc0eae9f3b7aa00edabc6ea9306"} Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.991106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5mrnk" event={"ID":"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1","Type":"ContainerStarted","Data":"ceb9e9fdc0fbe0e01c1cff1e55fde32653535576aeb494e4e54de204faf484cd"} Dec 04 14:18:29 crc kubenswrapper[4715]: I1204 14:18:29.998706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" event={"ID":"5cc0cd44-94c5-41c6-a11b-06faa5c78299","Type":"ContainerStarted","Data":"e10be484f5fc1865590c6a9d09aeaffcbb8e8ddc3c8828123235abc825139d7b"} Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.253130 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.254903 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.261245 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.261454 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-886p6" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.264539 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.286673 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z4zqf"] Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.305675 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:18:30 crc kubenswrapper[4715]: W1204 14:18:30.313688 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e56f804_b6ae_4d86_8e0b_9aed08a299b4.slice/crio-88b8d4f78d7dc50dd59ee57325750e751996ee854cdcef8f39f0cf8fc3d1e6c8 WatchSource:0}: Error finding container 88b8d4f78d7dc50dd59ee57325750e751996ee854cdcef8f39f0cf8fc3d1e6c8: Status 404 returned error can't find the container with id 88b8d4f78d7dc50dd59ee57325750e751996ee854cdcef8f39f0cf8fc3d1e6c8 Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.451292 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.451332 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-config-data\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.451691 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qph85\" (UniqueName: \"kubernetes.io/projected/90345add-0cc7-4bb9-879e-cd38508aa7e4-kube-api-access-qph85\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.451737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.452182 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-logs\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.452235 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.452254 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-scripts\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.518838 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.522551 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.525870 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.544466 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.553911 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.553958 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-config-data\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.554067 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qph85\" (UniqueName: \"kubernetes.io/projected/90345add-0cc7-4bb9-879e-cd38508aa7e4-kube-api-access-qph85\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.554103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.554159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-logs\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.554192 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.554258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-scripts\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.556706 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-logs\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.556775 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.557119 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.561889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-config-data\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.571007 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-scripts\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.572117 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.579845 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qph85\" (UniqueName: \"kubernetes.io/projected/90345add-0cc7-4bb9-879e-cd38508aa7e4-kube-api-access-qph85\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.603279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " pod="openstack/glance-default-external-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.655958 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.656056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.656084 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.656124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.656210 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-logs\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.656243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.656296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx5vt\" (UniqueName: \"kubernetes.io/projected/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-kube-api-access-kx5vt\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.757987 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-logs\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758111 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx5vt\" (UniqueName: \"kubernetes.io/projected/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-kube-api-access-kx5vt\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758312 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758339 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758393 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758657 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-logs\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.758962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.759117 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.767879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.767987 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.769850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.786295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx5vt\" (UniqueName: \"kubernetes.io/projected/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-kube-api-access-kx5vt\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.788473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.848286 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:18:30 crc kubenswrapper[4715]: I1204 14:18:30.880779 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.011072 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" event={"ID":"4e56f804-b6ae-4d86-8e0b-9aed08a299b4","Type":"ContainerStarted","Data":"8f255ea6ba69fe26492e6e42a0adcf97b1c6441c697666999c812155e5eb0e6d"} Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.011410 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" event={"ID":"4e56f804-b6ae-4d86-8e0b-9aed08a299b4","Type":"ContainerStarted","Data":"88b8d4f78d7dc50dd59ee57325750e751996ee854cdcef8f39f0cf8fc3d1e6c8"} Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.014649 4715 generic.go:334] "Generic (PLEG): container finished" podID="5cc0cd44-94c5-41c6-a11b-06faa5c78299" containerID="e10be484f5fc1865590c6a9d09aeaffcbb8e8ddc3c8828123235abc825139d7b" exitCode=0 Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.014762 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" event={"ID":"5cc0cd44-94c5-41c6-a11b-06faa5c78299","Type":"ContainerDied","Data":"e10be484f5fc1865590c6a9d09aeaffcbb8e8ddc3c8828123235abc825139d7b"} Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.406385 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-5mrnk" podStartSLOduration=7.406358476 podStartE2EDuration="7.406358476s" podCreationTimestamp="2025-12-04 14:18:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:31.37070689 +0000 UTC m=+1268.439425115" watchObservedRunningTime="2025-12-04 14:18:31.406358476 +0000 UTC m=+1268.475076701" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.516760 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.571358 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:18:31 crc kubenswrapper[4715]: W1204 14:18:31.577472 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45e3d38b_7a18_4869_a7fe_a9a34acec6b4.slice/crio-c8a0b85f2507faba871b277c73e108074ae52777a223c0ed6184a5f343a9b97a WatchSource:0}: Error finding container c8a0b85f2507faba871b277c73e108074ae52777a223c0ed6184a5f343a9b97a: Status 404 returned error can't find the container with id c8a0b85f2507faba871b277c73e108074ae52777a223c0ed6184a5f343a9b97a Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.635303 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.679790 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-sb\") pod \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.679878 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-nb\") pod \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.679933 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs8gn\" (UniqueName: \"kubernetes.io/projected/5cc0cd44-94c5-41c6-a11b-06faa5c78299-kube-api-access-cs8gn\") pod \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.680023 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-config\") pod \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.680074 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-svc\") pod \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.680428 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-swift-storage-0\") pod \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\" (UID: \"5cc0cd44-94c5-41c6-a11b-06faa5c78299\") " Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.686274 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc0cd44-94c5-41c6-a11b-06faa5c78299-kube-api-access-cs8gn" (OuterVolumeSpecName: "kube-api-access-cs8gn") pod "5cc0cd44-94c5-41c6-a11b-06faa5c78299" (UID: "5cc0cd44-94c5-41c6-a11b-06faa5c78299"). InnerVolumeSpecName "kube-api-access-cs8gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.711842 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5cc0cd44-94c5-41c6-a11b-06faa5c78299" (UID: "5cc0cd44-94c5-41c6-a11b-06faa5c78299"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.713401 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5cc0cd44-94c5-41c6-a11b-06faa5c78299" (UID: "5cc0cd44-94c5-41c6-a11b-06faa5c78299"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.721848 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5cc0cd44-94c5-41c6-a11b-06faa5c78299" (UID: "5cc0cd44-94c5-41c6-a11b-06faa5c78299"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.722347 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-config" (OuterVolumeSpecName: "config") pod "5cc0cd44-94c5-41c6-a11b-06faa5c78299" (UID: "5cc0cd44-94c5-41c6-a11b-06faa5c78299"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.734894 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5cc0cd44-94c5-41c6-a11b-06faa5c78299" (UID: "5cc0cd44-94c5-41c6-a11b-06faa5c78299"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.786550 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.786594 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.786607 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs8gn\" (UniqueName: \"kubernetes.io/projected/5cc0cd44-94c5-41c6-a11b-06faa5c78299-kube-api-access-cs8gn\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.786619 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.786628 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:31 crc kubenswrapper[4715]: I1204 14:18:31.786636 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc0cd44-94c5-41c6-a11b-06faa5c78299-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.025247 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" event={"ID":"5cc0cd44-94c5-41c6-a11b-06faa5c78299","Type":"ContainerDied","Data":"ba31ffd36339fcaf717149af462289bd56a07d23a098e0e0e2515c629086adde"} Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.025596 4715 scope.go:117] "RemoveContainer" containerID="e10be484f5fc1865590c6a9d09aeaffcbb8e8ddc3c8828123235abc825139d7b" Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.025325 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-rjgzw" Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.026607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90345add-0cc7-4bb9-879e-cd38508aa7e4","Type":"ContainerStarted","Data":"863763aa97b361baa8b4bd72422412edf311ed128f82caa7d2a2d002b8c095b5"} Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.027654 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"45e3d38b-7a18-4869-a7fe-a9a34acec6b4","Type":"ContainerStarted","Data":"c8a0b85f2507faba871b277c73e108074ae52777a223c0ed6184a5f343a9b97a"} Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.030058 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerID="8f255ea6ba69fe26492e6e42a0adcf97b1c6441c697666999c812155e5eb0e6d" exitCode=0 Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.030098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" event={"ID":"4e56f804-b6ae-4d86-8e0b-9aed08a299b4","Type":"ContainerDied","Data":"8f255ea6ba69fe26492e6e42a0adcf97b1c6441c697666999c812155e5eb0e6d"} Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.096106 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-rjgzw"] Dec 04 14:18:32 crc kubenswrapper[4715]: I1204 14:18:32.103347 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-rjgzw"] Dec 04 14:18:33 crc kubenswrapper[4715]: I1204 14:18:33.206217 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc0cd44-94c5-41c6-a11b-06faa5c78299" path="/var/lib/kubelet/pods/5cc0cd44-94c5-41c6-a11b-06faa5c78299/volumes" Dec 04 14:18:36 crc kubenswrapper[4715]: I1204 14:18:36.846923 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:18:36 crc kubenswrapper[4715]: I1204 14:18:36.942160 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.197619 4715 generic.go:334] "Generic (PLEG): container finished" podID="8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" containerID="d54ffff1c0ecf1f8e43275f596d60721c329f39ceab316eaa8dde2aa18974237" exitCode=0 Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.197675 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z2s4m" event={"ID":"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253","Type":"ContainerDied","Data":"d54ffff1c0ecf1f8e43275f596d60721c329f39ceab316eaa8dde2aa18974237"} Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.260998 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8599dd67c5-qs56p"] Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.315771 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-754dbdc69b-nwpzv"] Dec 04 14:18:38 crc kubenswrapper[4715]: E1204 14:18:38.321373 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc0cd44-94c5-41c6-a11b-06faa5c78299" containerName="init" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.321403 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc0cd44-94c5-41c6-a11b-06faa5c78299" containerName="init" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.321745 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc0cd44-94c5-41c6-a11b-06faa5c78299" containerName="init" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.322992 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.331232 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378176 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-config-data\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378224 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-scripts\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378248 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88520da5-8fee-4a16-9396-2280c468c41a-logs\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378295 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-combined-ca-bundle\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378344 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpp92\" (UniqueName: \"kubernetes.io/projected/88520da5-8fee-4a16-9396-2280c468c41a-kube-api-access-xpp92\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378361 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-tls-certs\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.378378 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-secret-key\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.385618 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-754dbdc69b-nwpzv"] Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.449421 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57458fcb8f-w827q"] Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-combined-ca-bundle\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpp92\" (UniqueName: \"kubernetes.io/projected/88520da5-8fee-4a16-9396-2280c468c41a-kube-api-access-xpp92\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480667 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-tls-certs\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480682 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-secret-key\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480750 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-config-data\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480776 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-scripts\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.480801 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88520da5-8fee-4a16-9396-2280c468c41a-logs\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.482015 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88520da5-8fee-4a16-9396-2280c468c41a-logs\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.487946 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-config-data\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.493164 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-scripts\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.501858 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-secret-key\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.509835 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b798c69c6-g6fvm"] Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.511458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.511953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-combined-ca-bundle\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.513562 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpp92\" (UniqueName: \"kubernetes.io/projected/88520da5-8fee-4a16-9396-2280c468c41a-kube-api-access-xpp92\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.513770 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-tls-certs\") pod \"horizon-754dbdc69b-nwpzv\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.523061 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b798c69c6-g6fvm"] Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.684425 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-horizon-secret-key\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.685196 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-combined-ca-bundle\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.685277 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8552f780-12e2-4eeb-bbc5-9ef637e9278f-config-data\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.685331 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w2ks\" (UniqueName: \"kubernetes.io/projected/8552f780-12e2-4eeb-bbc5-9ef637e9278f-kube-api-access-7w2ks\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.685362 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-horizon-tls-certs\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.685418 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8552f780-12e2-4eeb-bbc5-9ef637e9278f-scripts\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.685500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8552f780-12e2-4eeb-bbc5-9ef637e9278f-logs\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.758342 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.758403 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.758448 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.759226 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d601d4f16680a55d11b4648fd80ee11fd0f65f95595398da0f63f96f86a77a1f"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.759287 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://d601d4f16680a55d11b4648fd80ee11fd0f65f95595398da0f63f96f86a77a1f" gracePeriod=600 Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.769538 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.790838 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8552f780-12e2-4eeb-bbc5-9ef637e9278f-config-data\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.791013 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w2ks\" (UniqueName: \"kubernetes.io/projected/8552f780-12e2-4eeb-bbc5-9ef637e9278f-kube-api-access-7w2ks\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.791186 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-horizon-tls-certs\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.792569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8552f780-12e2-4eeb-bbc5-9ef637e9278f-config-data\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.793254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8552f780-12e2-4eeb-bbc5-9ef637e9278f-scripts\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.793376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8552f780-12e2-4eeb-bbc5-9ef637e9278f-logs\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.794378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8552f780-12e2-4eeb-bbc5-9ef637e9278f-logs\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.794552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-horizon-secret-key\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.794647 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-combined-ca-bundle\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.794905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8552f780-12e2-4eeb-bbc5-9ef637e9278f-scripts\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.803580 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-horizon-secret-key\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.803797 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-horizon-tls-certs\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.807899 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8552f780-12e2-4eeb-bbc5-9ef637e9278f-combined-ca-bundle\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.808067 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w2ks\" (UniqueName: \"kubernetes.io/projected/8552f780-12e2-4eeb-bbc5-9ef637e9278f-kube-api-access-7w2ks\") pod \"horizon-5b798c69c6-g6fvm\" (UID: \"8552f780-12e2-4eeb-bbc5-9ef637e9278f\") " pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:38 crc kubenswrapper[4715]: I1204 14:18:38.882989 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:18:39 crc kubenswrapper[4715]: I1204 14:18:39.211408 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="d601d4f16680a55d11b4648fd80ee11fd0f65f95595398da0f63f96f86a77a1f" exitCode=0 Dec 04 14:18:39 crc kubenswrapper[4715]: I1204 14:18:39.211465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"d601d4f16680a55d11b4648fd80ee11fd0f65f95595398da0f63f96f86a77a1f"} Dec 04 14:18:39 crc kubenswrapper[4715]: I1204 14:18:39.214517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"45e3d38b-7a18-4869-a7fe-a9a34acec6b4","Type":"ContainerStarted","Data":"6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0"} Dec 04 14:18:41 crc kubenswrapper[4715]: E1204 14:18:41.224455 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 04 14:18:41 crc kubenswrapper[4715]: E1204 14:18:41.225100 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8g8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-8bbsj_openstack(2ddb5944-20e2-4f19-96c4-ee93ac8cfa02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:18:41 crc kubenswrapper[4715]: E1204 14:18:41.226259 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-8bbsj" podUID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" Dec 04 14:18:41 crc kubenswrapper[4715]: E1204 14:18:41.239793 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-8bbsj" podUID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.008654 4715 scope.go:117] "RemoveContainer" containerID="60dea1351c777fadc6f4bbc06b373dac4289c7f94189506529170a4446c48e46" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.105644 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.214493 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-credential-keys\") pod \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.214605 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l25vj\" (UniqueName: \"kubernetes.io/projected/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-kube-api-access-l25vj\") pod \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.214644 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-combined-ca-bundle\") pod \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.214674 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-fernet-keys\") pod \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.214854 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-scripts\") pod \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.214917 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-config-data\") pod \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\" (UID: \"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253\") " Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.220238 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" (UID: "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.220462 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" (UID: "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.232336 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-kube-api-access-l25vj" (OuterVolumeSpecName: "kube-api-access-l25vj") pod "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" (UID: "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253"). InnerVolumeSpecName "kube-api-access-l25vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.237798 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-scripts" (OuterVolumeSpecName: "scripts") pod "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" (UID: "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.247696 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" (UID: "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.255626 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-config-data" (OuterVolumeSpecName: "config-data") pod "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" (UID: "8d2c3b9d-ffd1-46db-a3f5-4f113daa2253"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.317318 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.317354 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.317371 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.317388 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l25vj\" (UniqueName: \"kubernetes.io/projected/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-kube-api-access-l25vj\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.317402 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.317413 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.325399 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z2s4m" event={"ID":"8d2c3b9d-ffd1-46db-a3f5-4f113daa2253","Type":"ContainerDied","Data":"83c91bdca9cec1ebd4f70cbee8e003cb61e4c0f731b8c85ea9a557a256c5a6e2"} Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.325446 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83c91bdca9cec1ebd4f70cbee8e003cb61e4c0f731b8c85ea9a557a256c5a6e2" Dec 04 14:18:50 crc kubenswrapper[4715]: I1204 14:18:50.325450 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z2s4m" Dec 04 14:18:50 crc kubenswrapper[4715]: E1204 14:18:50.454568 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d2c3b9d_ffd1_46db_a3f5_4f113daa2253.slice\": RecentStats: unable to find data in memory cache]" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.193614 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-z2s4m"] Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.203649 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-z2s4m"] Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.298154 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5pp8k"] Dec 04 14:18:51 crc kubenswrapper[4715]: E1204 14:18:51.298556 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" containerName="keystone-bootstrap" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.298572 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" containerName="keystone-bootstrap" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.298745 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" containerName="keystone-bootstrap" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.299367 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.301626 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.301632 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.301957 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bskrn" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.305181 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.308632 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5pp8k"] Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.315354 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:18:51 crc kubenswrapper[4715]: E1204 14:18:51.428676 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 04 14:18:51 crc kubenswrapper[4715]: E1204 14:18:51.428855 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2jqbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-cgdwm_openstack(2ca3112c-c14e-41af-ad25-c16c7ca1a8c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:18:51 crc kubenswrapper[4715]: E1204 14:18:51.431245 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-cgdwm" podUID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.439111 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m7pt\" (UniqueName: \"kubernetes.io/projected/fb889380-79fe-461a-9a87-85c13fd8c716-kube-api-access-4m7pt\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.439187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-credential-keys\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.439313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-combined-ca-bundle\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.439483 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-scripts\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.439608 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-config-data\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.439683 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-fernet-keys\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.541411 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-config-data\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.541485 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-fernet-keys\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.541533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m7pt\" (UniqueName: \"kubernetes.io/projected/fb889380-79fe-461a-9a87-85c13fd8c716-kube-api-access-4m7pt\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.541555 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-credential-keys\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.541632 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-combined-ca-bundle\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.541694 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-scripts\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.547319 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-fernet-keys\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.547324 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-combined-ca-bundle\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.547604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-config-data\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.548335 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-credential-keys\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.548395 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-scripts\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.559547 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m7pt\" (UniqueName: \"kubernetes.io/projected/fb889380-79fe-461a-9a87-85c13fd8c716-kube-api-access-4m7pt\") pod \"keystone-bootstrap-5pp8k\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:51 crc kubenswrapper[4715]: I1204 14:18:51.622739 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:18:52 crc kubenswrapper[4715]: E1204 14:18:52.343490 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-cgdwm" podUID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" Dec 04 14:18:53 crc kubenswrapper[4715]: E1204 14:18:53.175752 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 04 14:18:53 crc kubenswrapper[4715]: E1204 14:18:53.176343 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nh7fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-s5j8r_openstack(7e311957-d00b-43da-a9dc-ca48386dfa88): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:18:53 crc kubenswrapper[4715]: E1204 14:18:53.178212 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-s5j8r" podUID="7e311957-d00b-43da-a9dc-ca48386dfa88" Dec 04 14:18:53 crc kubenswrapper[4715]: I1204 14:18:53.193639 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2c3b9d-ffd1-46db-a3f5-4f113daa2253" path="/var/lib/kubelet/pods/8d2c3b9d-ffd1-46db-a3f5-4f113daa2253/volumes" Dec 04 14:18:53 crc kubenswrapper[4715]: E1204 14:18:53.349855 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-s5j8r" podUID="7e311957-d00b-43da-a9dc-ca48386dfa88" Dec 04 14:18:55 crc kubenswrapper[4715]: I1204 14:18:55.293664 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-754dbdc69b-nwpzv"] Dec 04 14:18:55 crc kubenswrapper[4715]: I1204 14:18:55.346742 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b798c69c6-g6fvm"] Dec 04 14:18:58 crc kubenswrapper[4715]: W1204 14:18:58.030294 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8552f780_12e2_4eeb_bbc5_9ef637e9278f.slice/crio-8804424cf0560362c3f40b77235a308268ced133b1f4642a0e4d49dcee62e09d WatchSource:0}: Error finding container 8804424cf0560362c3f40b77235a308268ced133b1f4642a0e4d49dcee62e09d: Status 404 returned error can't find the container with id 8804424cf0560362c3f40b77235a308268ced133b1f4642a0e4d49dcee62e09d Dec 04 14:18:58 crc kubenswrapper[4715]: W1204 14:18:58.032102 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88520da5_8fee_4a16_9396_2280c468c41a.slice/crio-9b8b7a0b24c121328511b9162416397e6e369f78ee03d2c74f62269c993848c2 WatchSource:0}: Error finding container 9b8b7a0b24c121328511b9162416397e6e369f78ee03d2c74f62269c993848c2: Status 404 returned error can't find the container with id 9b8b7a0b24c121328511b9162416397e6e369f78ee03d2c74f62269c993848c2 Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.406321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"5f8cf4969e2a565cf9483d0133bd3e156b294d54b874a31409400fd261c14615"} Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.408127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754dbdc69b-nwpzv" event={"ID":"88520da5-8fee-4a16-9396-2280c468c41a","Type":"ContainerStarted","Data":"9b8b7a0b24c121328511b9162416397e6e369f78ee03d2c74f62269c993848c2"} Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.412140 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b798c69c6-g6fvm" event={"ID":"8552f780-12e2-4eeb-bbc5-9ef637e9278f","Type":"ContainerStarted","Data":"8804424cf0560362c3f40b77235a308268ced133b1f4642a0e4d49dcee62e09d"} Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.416115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" event={"ID":"4e56f804-b6ae-4d86-8e0b-9aed08a299b4","Type":"ContainerStarted","Data":"bdd78e6d8ce2e391a2a3c679b069dc2b9e4dc37f274bf1dbe8a52828ada6e4b2"} Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.416562 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.456911 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" podStartSLOduration=29.456890179 podStartE2EDuration="29.456890179s" podCreationTimestamp="2025-12-04 14:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:58.448536193 +0000 UTC m=+1295.517254408" watchObservedRunningTime="2025-12-04 14:18:58.456890179 +0000 UTC m=+1295.525608394" Dec 04 14:18:58 crc kubenswrapper[4715]: I1204 14:18:58.826115 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5pp8k"] Dec 04 14:18:58 crc kubenswrapper[4715]: W1204 14:18:58.853550 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb889380_79fe_461a_9a87_85c13fd8c716.slice/crio-ffd9601ebf78eed72e9d717c3e4a19ecfe5366785e01577ae468e496c024b210 WatchSource:0}: Error finding container ffd9601ebf78eed72e9d717c3e4a19ecfe5366785e01577ae468e496c024b210: Status 404 returned error can't find the container with id ffd9601ebf78eed72e9d717c3e4a19ecfe5366785e01577ae468e496c024b210 Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.446074 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b798c69c6-g6fvm" event={"ID":"8552f780-12e2-4eeb-bbc5-9ef637e9278f","Type":"ContainerStarted","Data":"ef9703c93b640e58b99296474b68500b37e5ff9b6c2f5aee282c68b7abd0fff5"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.448591 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pp8k" event={"ID":"fb889380-79fe-461a-9a87-85c13fd8c716","Type":"ContainerStarted","Data":"ffd9601ebf78eed72e9d717c3e4a19ecfe5366785e01577ae468e496c024b210"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.452877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8bbsj" event={"ID":"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02","Type":"ContainerStarted","Data":"f72345280247cf84a4862946e0d8c5962be25d237d103214868392cdcc6cf55e"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.459319 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8599dd67c5-qs56p" event={"ID":"e8331eb7-653b-4a6c-ada8-f13dcf916a8a","Type":"ContainerStarted","Data":"dbb5250430bbee33218c769d14a9d9d05e1495c461ecb329525091ccaaed8b15"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.466538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57458fcb8f-w827q" event={"ID":"a1c0859d-f9ae-457f-9c44-ea9891d7756b","Type":"ContainerStarted","Data":"6c96e1ac3cfcae2163fd4c6676397b563df5edeb5e4005f1c1459da15c59f831"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.477590 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754dbdc69b-nwpzv" event={"ID":"88520da5-8fee-4a16-9396-2280c468c41a","Type":"ContainerStarted","Data":"1341f08949be200925226253e4ab03bfe80d9e1c59dcf0907a757e3dab2e4ec3"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.478579 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-8bbsj" podStartSLOduration=2.752541222 podStartE2EDuration="35.478562028s" podCreationTimestamp="2025-12-04 14:18:24 +0000 UTC" firstStartedPulling="2025-12-04 14:18:25.839966908 +0000 UTC m=+1262.908685123" lastFinishedPulling="2025-12-04 14:18:58.565987714 +0000 UTC m=+1295.634705929" observedRunningTime="2025-12-04 14:18:59.475738471 +0000 UTC m=+1296.544456686" watchObservedRunningTime="2025-12-04 14:18:59.478562028 +0000 UTC m=+1296.547280533" Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.484086 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerStarted","Data":"de12a58b72d05fca1fe302798225bc3ba3828024703f7b8f60ac4c0a898741c7"} Dec 04 14:18:59 crc kubenswrapper[4715]: I1204 14:18:59.486825 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5978d7d969-7vwfh" event={"ID":"0a909008-f754-49fb-87cb-dc0d09c91bfd","Type":"ContainerStarted","Data":"80bb690535d08957f268fb451a9b2608f1e5529fbe003046329fd5dbf9c7373b"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.509490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"45e3d38b-7a18-4869-a7fe-a9a34acec6b4","Type":"ContainerStarted","Data":"b81f2735ecdfa1ac2509cc007f5c75bbfff2f5602b4a7dd01b7f7a9c0dd8069b"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.509652 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-log" containerID="cri-o://6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.510331 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-httpd" containerID="cri-o://b81f2735ecdfa1ac2509cc007f5c75bbfff2f5602b4a7dd01b7f7a9c0dd8069b" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.519345 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8599dd67c5-qs56p" event={"ID":"e8331eb7-653b-4a6c-ada8-f13dcf916a8a","Type":"ContainerStarted","Data":"207bb21cc12f37dde2b5a2635bd91b784400406326add9b1bbfa01c6fe3a1eb4"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.519598 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8599dd67c5-qs56p" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon-log" containerID="cri-o://dbb5250430bbee33218c769d14a9d9d05e1495c461ecb329525091ccaaed8b15" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.521156 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8599dd67c5-qs56p" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon" containerID="cri-o://207bb21cc12f37dde2b5a2635bd91b784400406326add9b1bbfa01c6fe3a1eb4" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.528797 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57458fcb8f-w827q" event={"ID":"a1c0859d-f9ae-457f-9c44-ea9891d7756b","Type":"ContainerStarted","Data":"cde5fc09e99af07fca1b14c1394b5942eed07a676cad8a6acb20ea77d12058e6"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.528978 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57458fcb8f-w827q" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon-log" containerID="cri-o://6c96e1ac3cfcae2163fd4c6676397b563df5edeb5e4005f1c1459da15c59f831" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.529097 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57458fcb8f-w827q" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon" containerID="cri-o://cde5fc09e99af07fca1b14c1394b5942eed07a676cad8a6acb20ea77d12058e6" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.536603 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754dbdc69b-nwpzv" event={"ID":"88520da5-8fee-4a16-9396-2280c468c41a","Type":"ContainerStarted","Data":"f1297a46867ceb825461ae2c754dc4d5b3ba962c40fa5376fd6f94ccf7879eb0"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.546793 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5978d7d969-7vwfh" event={"ID":"0a909008-f754-49fb-87cb-dc0d09c91bfd","Type":"ContainerStarted","Data":"d9d5504dd22df7cd3fddb8b08067bef9d4f4ee6732924b53b10b4c34fa598d9e"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.546869 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5978d7d969-7vwfh" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon-log" containerID="cri-o://80bb690535d08957f268fb451a9b2608f1e5529fbe003046329fd5dbf9c7373b" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.546890 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5978d7d969-7vwfh" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon" containerID="cri-o://d9d5504dd22df7cd3fddb8b08067bef9d4f4ee6732924b53b10b4c34fa598d9e" gracePeriod=30 Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.560506 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=31.560477978 podStartE2EDuration="31.560477978s" podCreationTimestamp="2025-12-04 14:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:00.54615923 +0000 UTC m=+1297.614877465" watchObservedRunningTime="2025-12-04 14:19:00.560477978 +0000 UTC m=+1297.629196193" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.572210 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-57458fcb8f-w827q" podStartSLOduration=4.268084451 podStartE2EDuration="33.572184485s" podCreationTimestamp="2025-12-04 14:18:27 +0000 UTC" firstStartedPulling="2025-12-04 14:18:28.983885441 +0000 UTC m=+1266.052603656" lastFinishedPulling="2025-12-04 14:18:58.287985475 +0000 UTC m=+1295.356703690" observedRunningTime="2025-12-04 14:19:00.569516253 +0000 UTC m=+1297.638234468" watchObservedRunningTime="2025-12-04 14:19:00.572184485 +0000 UTC m=+1297.640902690" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.590532 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b798c69c6-g6fvm" event={"ID":"8552f780-12e2-4eeb-bbc5-9ef637e9278f","Type":"ContainerStarted","Data":"daf792e29db6a3bc9401d93fb9dbf06e618fc740998a832ae915a5a7f7bbd0c5"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.594508 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-754dbdc69b-nwpzv" podStartSLOduration=22.594492329 podStartE2EDuration="22.594492329s" podCreationTimestamp="2025-12-04 14:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:00.593610106 +0000 UTC m=+1297.662328321" watchObservedRunningTime="2025-12-04 14:19:00.594492329 +0000 UTC m=+1297.663210544" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.607617 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pp8k" event={"ID":"fb889380-79fe-461a-9a87-85c13fd8c716","Type":"ContainerStarted","Data":"60669bee47edc974d542a61a67917981b56c1c8b8cb397beeba5befbdbfdda40"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.614843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90345add-0cc7-4bb9-879e-cd38508aa7e4","Type":"ContainerStarted","Data":"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de"} Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.665890 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8599dd67c5-qs56p" podStartSLOduration=6.302725118 podStartE2EDuration="36.665855972s" podCreationTimestamp="2025-12-04 14:18:24 +0000 UTC" firstStartedPulling="2025-12-04 14:18:28.026321789 +0000 UTC m=+1265.095040004" lastFinishedPulling="2025-12-04 14:18:58.389452643 +0000 UTC m=+1295.458170858" observedRunningTime="2025-12-04 14:19:00.619443025 +0000 UTC m=+1297.688161240" watchObservedRunningTime="2025-12-04 14:19:00.665855972 +0000 UTC m=+1297.734574297" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.666942 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5978d7d969-7vwfh" podStartSLOduration=9.163468903 podStartE2EDuration="36.666930941s" podCreationTimestamp="2025-12-04 14:18:24 +0000 UTC" firstStartedPulling="2025-12-04 14:18:25.537790885 +0000 UTC m=+1262.606509100" lastFinishedPulling="2025-12-04 14:18:53.041252923 +0000 UTC m=+1290.109971138" observedRunningTime="2025-12-04 14:19:00.645184402 +0000 UTC m=+1297.713902627" watchObservedRunningTime="2025-12-04 14:19:00.666930941 +0000 UTC m=+1297.735649156" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.689756 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5b798c69c6-g6fvm" podStartSLOduration=22.689736459 podStartE2EDuration="22.689736459s" podCreationTimestamp="2025-12-04 14:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:00.67094468 +0000 UTC m=+1297.739662895" watchObservedRunningTime="2025-12-04 14:19:00.689736459 +0000 UTC m=+1297.758454674" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.699888 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5pp8k" podStartSLOduration=9.699862823 podStartE2EDuration="9.699862823s" podCreationTimestamp="2025-12-04 14:18:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:00.69346288 +0000 UTC m=+1297.762181105" watchObservedRunningTime="2025-12-04 14:19:00.699862823 +0000 UTC m=+1297.768581048" Dec 04 14:19:00 crc kubenswrapper[4715]: E1204 14:19:00.770237 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45e3d38b_7a18_4869_a7fe_a9a34acec6b4.slice/crio-6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45e3d38b_7a18_4869_a7fe_a9a34acec6b4.slice/crio-conmon-6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.849353 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:00 crc kubenswrapper[4715]: I1204 14:19:00.849553 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.625756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90345add-0cc7-4bb9-879e-cd38508aa7e4","Type":"ContainerStarted","Data":"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b"} Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.625833 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-log" containerID="cri-o://614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de" gracePeriod=30 Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.626112 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-httpd" containerID="cri-o://c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b" gracePeriod=30 Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.634485 4715 generic.go:334] "Generic (PLEG): container finished" podID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerID="b81f2735ecdfa1ac2509cc007f5c75bbfff2f5602b4a7dd01b7f7a9c0dd8069b" exitCode=0 Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.634521 4715 generic.go:334] "Generic (PLEG): container finished" podID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerID="6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0" exitCode=143 Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.634618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"45e3d38b-7a18-4869-a7fe-a9a34acec6b4","Type":"ContainerDied","Data":"b81f2735ecdfa1ac2509cc007f5c75bbfff2f5602b4a7dd01b7f7a9c0dd8069b"} Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.634680 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"45e3d38b-7a18-4869-a7fe-a9a34acec6b4","Type":"ContainerDied","Data":"6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0"} Dec 04 14:19:01 crc kubenswrapper[4715]: I1204 14:19:01.660174 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=32.660154669 podStartE2EDuration="32.660154669s" podCreationTimestamp="2025-12-04 14:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:01.648800621 +0000 UTC m=+1298.717518836" watchObservedRunningTime="2025-12-04 14:19:01.660154669 +0000 UTC m=+1298.728872884" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.086542 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178359 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx5vt\" (UniqueName: \"kubernetes.io/projected/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-kube-api-access-kx5vt\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178661 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-logs\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178771 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-combined-ca-bundle\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178874 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-config-data\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178943 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-scripts\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178963 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.178999 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-httpd-run\") pod \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\" (UID: \"45e3d38b-7a18-4869-a7fe-a9a34acec6b4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.179048 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-logs" (OuterVolumeSpecName: "logs") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.179966 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.184135 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.184403 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.186728 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-scripts" (OuterVolumeSpecName: "scripts") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.190747 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-kube-api-access-kx5vt" (OuterVolumeSpecName: "kube-api-access-kx5vt") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "kube-api-access-kx5vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.277644 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-config-data" (OuterVolumeSpecName: "config-data") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.282298 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.282520 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.282559 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.282574 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.282586 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx5vt\" (UniqueName: \"kubernetes.io/projected/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-kube-api-access-kx5vt\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.294418 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45e3d38b-7a18-4869-a7fe-a9a34acec6b4" (UID: "45e3d38b-7a18-4869-a7fe-a9a34acec6b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.308370 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.348481 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383307 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-combined-ca-bundle\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383416 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-httpd-run\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383462 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-config-data\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383487 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383524 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-logs\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383547 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qph85\" (UniqueName: \"kubernetes.io/projected/90345add-0cc7-4bb9-879e-cd38508aa7e4-kube-api-access-qph85\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.383638 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-scripts\") pod \"90345add-0cc7-4bb9-879e-cd38508aa7e4\" (UID: \"90345add-0cc7-4bb9-879e-cd38508aa7e4\") " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.384010 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e3d38b-7a18-4869-a7fe-a9a34acec6b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.384046 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.384048 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.389104 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-logs" (OuterVolumeSpecName: "logs") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.389672 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-scripts" (OuterVolumeSpecName: "scripts") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.389673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.390169 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90345add-0cc7-4bb9-879e-cd38508aa7e4-kube-api-access-qph85" (OuterVolumeSpecName: "kube-api-access-qph85") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "kube-api-access-qph85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.417239 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.475946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-config-data" (OuterVolumeSpecName: "config-data") pod "90345add-0cc7-4bb9-879e-cd38508aa7e4" (UID: "90345add-0cc7-4bb9-879e-cd38508aa7e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485080 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485114 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485140 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485151 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90345add-0cc7-4bb9-879e-cd38508aa7e4-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485162 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qph85\" (UniqueName: \"kubernetes.io/projected/90345add-0cc7-4bb9-879e-cd38508aa7e4-kube-api-access-qph85\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485174 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.485183 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90345add-0cc7-4bb9-879e-cd38508aa7e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.535068 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.590484 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657207 4715 generic.go:334] "Generic (PLEG): container finished" podID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerID="c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b" exitCode=0 Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657247 4715 generic.go:334] "Generic (PLEG): container finished" podID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerID="614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de" exitCode=143 Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657305 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90345add-0cc7-4bb9-879e-cd38508aa7e4","Type":"ContainerDied","Data":"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b"} Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657338 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90345add-0cc7-4bb9-879e-cd38508aa7e4","Type":"ContainerDied","Data":"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de"} Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90345add-0cc7-4bb9-879e-cd38508aa7e4","Type":"ContainerDied","Data":"863763aa97b361baa8b4bd72422412edf311ed128f82caa7d2a2d002b8c095b5"} Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657368 4715 scope.go:117] "RemoveContainer" containerID="c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.657515 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.669728 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"45e3d38b-7a18-4869-a7fe-a9a34acec6b4","Type":"ContainerDied","Data":"c8a0b85f2507faba871b277c73e108074ae52777a223c0ed6184a5f343a9b97a"} Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.669829 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.689314 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerStarted","Data":"784803b676bc40f74c8cb814c6ded8ae84d6a7c073a2a93c40de59bd325ab990"} Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.702303 4715 scope.go:117] "RemoveContainer" containerID="614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.720763 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.742107 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.743654 4715 scope.go:117] "RemoveContainer" containerID="c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b" Dec 04 14:19:02 crc kubenswrapper[4715]: E1204 14:19:02.745668 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b\": container with ID starting with c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b not found: ID does not exist" containerID="c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.745712 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b"} err="failed to get container status \"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b\": rpc error: code = NotFound desc = could not find container \"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b\": container with ID starting with c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b not found: ID does not exist" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.745742 4715 scope.go:117] "RemoveContainer" containerID="614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de" Dec 04 14:19:02 crc kubenswrapper[4715]: E1204 14:19:02.749166 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de\": container with ID starting with 614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de not found: ID does not exist" containerID="614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.749226 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de"} err="failed to get container status \"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de\": rpc error: code = NotFound desc = could not find container \"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de\": container with ID starting with 614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de not found: ID does not exist" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.749264 4715 scope.go:117] "RemoveContainer" containerID="c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.753150 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b"} err="failed to get container status \"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b\": rpc error: code = NotFound desc = could not find container \"c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b\": container with ID starting with c216541600911c45967ef0a64d28858b8eb3e85d7602009e5ec0dc1071666a8b not found: ID does not exist" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.753190 4715 scope.go:117] "RemoveContainer" containerID="614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.756813 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de"} err="failed to get container status \"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de\": rpc error: code = NotFound desc = could not find container \"614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de\": container with ID starting with 614663d37e4ac3db09e7d8e0a11b6ae76fc1c59d408e069ea5c43063920d04de not found: ID does not exist" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.756843 4715 scope.go:117] "RemoveContainer" containerID="b81f2735ecdfa1ac2509cc007f5c75bbfff2f5602b4a7dd01b7f7a9c0dd8069b" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.756946 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778268 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: E1204 14:19:02.778706 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-log" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778720 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-log" Dec 04 14:19:02 crc kubenswrapper[4715]: E1204 14:19:02.778735 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-httpd" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778741 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-httpd" Dec 04 14:19:02 crc kubenswrapper[4715]: E1204 14:19:02.778753 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-log" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778758 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-log" Dec 04 14:19:02 crc kubenswrapper[4715]: E1204 14:19:02.778787 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-httpd" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778792 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-httpd" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778959 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-log" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778981 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-httpd" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.778990 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" containerName="glance-log" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.779006 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" containerName="glance-httpd" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.779969 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.783829 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.784246 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.784906 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.787369 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.785265 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-886p6" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.798944 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.818089 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.819853 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.830984 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.831100 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.831893 4715 scope.go:117] "RemoveContainer" containerID="6b107a44decea4421cff930561b065852dda6bd7c69a983047f7f4a0af17eab0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.835803 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.900993 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-config-data\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901214 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-logs\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901256 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp5x6\" (UniqueName: \"kubernetes.io/projected/7d95418a-f1f5-4676-903e-67b2fee4417c-kube-api-access-hp5x6\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901522 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dsh8\" (UniqueName: \"kubernetes.io/projected/edf7d465-56e6-4185-8ac7-bdbff1e10866-kube-api-access-7dsh8\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901559 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901640 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901713 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901753 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901875 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901918 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-scripts\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:02 crc kubenswrapper[4715]: I1204 14:19:02.901966 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp5x6\" (UniqueName: \"kubernetes.io/projected/7d95418a-f1f5-4676-903e-67b2fee4417c-kube-api-access-hp5x6\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004515 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dsh8\" (UniqueName: \"kubernetes.io/projected/edf7d465-56e6-4185-8ac7-bdbff1e10866-kube-api-access-7dsh8\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004542 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004649 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004712 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004804 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-scripts\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-config-data\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004926 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-logs\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.004973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.005369 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.007724 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.008149 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.009751 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.010205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-logs\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.011579 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.023455 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.026773 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.027944 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.028871 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-scripts\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.028943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.031286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.031962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.034926 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp5x6\" (UniqueName: \"kubernetes.io/projected/7d95418a-f1f5-4676-903e-67b2fee4417c-kube-api-access-hp5x6\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.038491 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-config-data\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.039329 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dsh8\" (UniqueName: \"kubernetes.io/projected/edf7d465-56e6-4185-8ac7-bdbff1e10866-kube-api-access-7dsh8\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.065783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.068204 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.166691 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.194672 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45e3d38b-7a18-4869-a7fe-a9a34acec6b4" path="/var/lib/kubelet/pods/45e3d38b-7a18-4869-a7fe-a9a34acec6b4/volumes" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.195489 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90345add-0cc7-4bb9-879e-cd38508aa7e4" path="/var/lib/kubelet/pods/90345add-0cc7-4bb9-879e-cd38508aa7e4/volumes" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.199732 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.677861 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:19:03 crc kubenswrapper[4715]: W1204 14:19:03.683155 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedf7d465_56e6_4185_8ac7_bdbff1e10866.slice/crio-6e497f9693dac055f4d3242f82848f44442908a01e68eea7344e1815df06829b WatchSource:0}: Error finding container 6e497f9693dac055f4d3242f82848f44442908a01e68eea7344e1815df06829b: Status 404 returned error can't find the container with id 6e497f9693dac055f4d3242f82848f44442908a01e68eea7344e1815df06829b Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.702576 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"edf7d465-56e6-4185-8ac7-bdbff1e10866","Type":"ContainerStarted","Data":"6e497f9693dac055f4d3242f82848f44442908a01e68eea7344e1815df06829b"} Dec 04 14:19:03 crc kubenswrapper[4715]: I1204 14:19:03.989722 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:19:04 crc kubenswrapper[4715]: W1204 14:19:04.003830 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d95418a_f1f5_4676_903e_67b2fee4417c.slice/crio-e3283855985c227b217b7a79ef3db5c1b02913560cf0d652b37702839d2a5481 WatchSource:0}: Error finding container e3283855985c227b217b7a79ef3db5c1b02913560cf0d652b37702839d2a5481: Status 404 returned error can't find the container with id e3283855985c227b217b7a79ef3db5c1b02913560cf0d652b37702839d2a5481 Dec 04 14:19:04 crc kubenswrapper[4715]: I1204 14:19:04.581834 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:19:04 crc kubenswrapper[4715]: I1204 14:19:04.729572 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:19:04 crc kubenswrapper[4715]: I1204 14:19:04.763287 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d95418a-f1f5-4676-903e-67b2fee4417c","Type":"ContainerStarted","Data":"e3283855985c227b217b7a79ef3db5c1b02913560cf0d652b37702839d2a5481"} Dec 04 14:19:04 crc kubenswrapper[4715]: I1204 14:19:04.798725 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tmg6v"] Dec 04 14:19:04 crc kubenswrapper[4715]: I1204 14:19:04.798934 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-tmg6v" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="dnsmasq-dns" containerID="cri-o://855aabe107d7d0ce74853d7a1bb68811a7a508c6ed41cc67583881d3d58fcfe9" gracePeriod=10 Dec 04 14:19:05 crc kubenswrapper[4715]: I1204 14:19:05.042991 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:19:06 crc kubenswrapper[4715]: I1204 14:19:06.805772 4715 generic.go:334] "Generic (PLEG): container finished" podID="fb889380-79fe-461a-9a87-85c13fd8c716" containerID="60669bee47edc974d542a61a67917981b56c1c8b8cb397beeba5befbdbfdda40" exitCode=0 Dec 04 14:19:06 crc kubenswrapper[4715]: I1204 14:19:06.806017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pp8k" event={"ID":"fb889380-79fe-461a-9a87-85c13fd8c716","Type":"ContainerDied","Data":"60669bee47edc974d542a61a67917981b56c1c8b8cb397beeba5befbdbfdda40"} Dec 04 14:19:06 crc kubenswrapper[4715]: I1204 14:19:06.814091 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"edf7d465-56e6-4185-8ac7-bdbff1e10866","Type":"ContainerStarted","Data":"c5b5d7b5a1ff4e61fcd49d9a1d37cf050a7d83db8680451c80b6e2e5269ec17e"} Dec 04 14:19:06 crc kubenswrapper[4715]: I1204 14:19:06.816148 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d95418a-f1f5-4676-903e-67b2fee4417c","Type":"ContainerStarted","Data":"34ebffad1ccd13f9dd6113bd6daea6476e89742be07724b139fd53e1c3592d0b"} Dec 04 14:19:06 crc kubenswrapper[4715]: I1204 14:19:06.818457 4715 generic.go:334] "Generic (PLEG): container finished" podID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerID="855aabe107d7d0ce74853d7a1bb68811a7a508c6ed41cc67583881d3d58fcfe9" exitCode=0 Dec 04 14:19:06 crc kubenswrapper[4715]: I1204 14:19:06.818496 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tmg6v" event={"ID":"332031a0-392b-4497-9c04-9c1bbd0a2ea0","Type":"ContainerDied","Data":"855aabe107d7d0ce74853d7a1bb68811a7a508c6ed41cc67583881d3d58fcfe9"} Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.104320 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.770339 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.770673 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.847724 4715 generic.go:334] "Generic (PLEG): container finished" podID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" containerID="f72345280247cf84a4862946e0d8c5962be25d237d103214868392cdcc6cf55e" exitCode=0 Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.847784 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8bbsj" event={"ID":"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02","Type":"ContainerDied","Data":"f72345280247cf84a4862946e0d8c5962be25d237d103214868392cdcc6cf55e"} Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.884223 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:19:08 crc kubenswrapper[4715]: I1204 14:19:08.884271 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.047901 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.051532 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.119929 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-config-data\") pod \"fb889380-79fe-461a-9a87-85c13fd8c716\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.119997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-combined-ca-bundle\") pod \"fb889380-79fe-461a-9a87-85c13fd8c716\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120091 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config\") pod \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-fernet-keys\") pod \"fb889380-79fe-461a-9a87-85c13fd8c716\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-dns-svc\") pod \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120191 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m7pt\" (UniqueName: \"kubernetes.io/projected/fb889380-79fe-461a-9a87-85c13fd8c716-kube-api-access-4m7pt\") pod \"fb889380-79fe-461a-9a87-85c13fd8c716\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120227 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fxzp\" (UniqueName: \"kubernetes.io/projected/332031a0-392b-4497-9c04-9c1bbd0a2ea0-kube-api-access-6fxzp\") pod \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120268 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-nb\") pod \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120357 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-scripts\") pod \"fb889380-79fe-461a-9a87-85c13fd8c716\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120398 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-credential-keys\") pod \"fb889380-79fe-461a-9a87-85c13fd8c716\" (UID: \"fb889380-79fe-461a-9a87-85c13fd8c716\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.120423 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-sb\") pod \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.128395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fb889380-79fe-461a-9a87-85c13fd8c716" (UID: "fb889380-79fe-461a-9a87-85c13fd8c716"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.143946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb889380-79fe-461a-9a87-85c13fd8c716-kube-api-access-4m7pt" (OuterVolumeSpecName: "kube-api-access-4m7pt") pod "fb889380-79fe-461a-9a87-85c13fd8c716" (UID: "fb889380-79fe-461a-9a87-85c13fd8c716"). InnerVolumeSpecName "kube-api-access-4m7pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.152907 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-scripts" (OuterVolumeSpecName: "scripts") pod "fb889380-79fe-461a-9a87-85c13fd8c716" (UID: "fb889380-79fe-461a-9a87-85c13fd8c716"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.152967 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332031a0-392b-4497-9c04-9c1bbd0a2ea0-kube-api-access-6fxzp" (OuterVolumeSpecName: "kube-api-access-6fxzp") pod "332031a0-392b-4497-9c04-9c1bbd0a2ea0" (UID: "332031a0-392b-4497-9c04-9c1bbd0a2ea0"). InnerVolumeSpecName "kube-api-access-6fxzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.188024 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fb889380-79fe-461a-9a87-85c13fd8c716" (UID: "fb889380-79fe-461a-9a87-85c13fd8c716"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.223546 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.223587 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m7pt\" (UniqueName: \"kubernetes.io/projected/fb889380-79fe-461a-9a87-85c13fd8c716-kube-api-access-4m7pt\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.223605 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fxzp\" (UniqueName: \"kubernetes.io/projected/332031a0-392b-4497-9c04-9c1bbd0a2ea0-kube-api-access-6fxzp\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.223619 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.223630 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.247087 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "332031a0-392b-4497-9c04-9c1bbd0a2ea0" (UID: "332031a0-392b-4497-9c04-9c1bbd0a2ea0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.259547 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-config-data" (OuterVolumeSpecName: "config-data") pod "fb889380-79fe-461a-9a87-85c13fd8c716" (UID: "fb889380-79fe-461a-9a87-85c13fd8c716"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.273327 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "332031a0-392b-4497-9c04-9c1bbd0a2ea0" (UID: "332031a0-392b-4497-9c04-9c1bbd0a2ea0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.276174 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "332031a0-392b-4497-9c04-9c1bbd0a2ea0" (UID: "332031a0-392b-4497-9c04-9c1bbd0a2ea0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.279986 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb889380-79fe-461a-9a87-85c13fd8c716" (UID: "fb889380-79fe-461a-9a87-85c13fd8c716"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.326575 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config" (OuterVolumeSpecName: "config") pod "332031a0-392b-4497-9c04-9c1bbd0a2ea0" (UID: "332031a0-392b-4497-9c04-9c1bbd0a2ea0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.327121 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config\") pod \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\" (UID: \"332031a0-392b-4497-9c04-9c1bbd0a2ea0\") " Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.327901 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.328003 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.328105 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.329325 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.329454 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb889380-79fe-461a-9a87-85c13fd8c716-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: W1204 14:19:11.327428 4715 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/332031a0-392b-4497-9c04-9c1bbd0a2ea0/volumes/kubernetes.io~configmap/config Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.329640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config" (OuterVolumeSpecName: "config") pod "332031a0-392b-4497-9c04-9c1bbd0a2ea0" (UID: "332031a0-392b-4497-9c04-9c1bbd0a2ea0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.431156 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332031a0-392b-4497-9c04-9c1bbd0a2ea0-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.909475 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tmg6v" event={"ID":"332031a0-392b-4497-9c04-9c1bbd0a2ea0","Type":"ContainerDied","Data":"2487bb0491b566f50604f4d77385a0b8ebb9a717da3a7a79331955e87581c950"} Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.910200 4715 scope.go:117] "RemoveContainer" containerID="855aabe107d7d0ce74853d7a1bb68811a7a508c6ed41cc67583881d3d58fcfe9" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.910366 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tmg6v" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.918636 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pp8k" event={"ID":"fb889380-79fe-461a-9a87-85c13fd8c716","Type":"ContainerDied","Data":"ffd9601ebf78eed72e9d717c3e4a19ecfe5366785e01577ae468e496c024b210"} Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.918677 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffd9601ebf78eed72e9d717c3e4a19ecfe5366785e01577ae468e496c024b210" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.918738 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pp8k" Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.946899 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"edf7d465-56e6-4185-8ac7-bdbff1e10866","Type":"ContainerStarted","Data":"74c00410246b6ff5e0221ad7de293d016e8bcf25cd908b648f975a4bc1b1aeba"} Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.954550 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d95418a-f1f5-4676-903e-67b2fee4417c","Type":"ContainerStarted","Data":"4e483519ce0acbca915c0fd7e57849d6f3637c62329e5d4be0e6008cdddc6593"} Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.968061 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tmg6v"] Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.988374 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tmg6v"] Dec 04 14:19:11 crc kubenswrapper[4715]: I1204 14:19:11.994634 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.994614222 podStartE2EDuration="9.994614222s" podCreationTimestamp="2025-12-04 14:19:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:11.977419367 +0000 UTC m=+1309.046137592" watchObservedRunningTime="2025-12-04 14:19:11.994614222 +0000 UTC m=+1309.063332437" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.027312 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.027288117 podStartE2EDuration="10.027288117s" podCreationTimestamp="2025-12-04 14:19:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:12.016827834 +0000 UTC m=+1309.085546049" watchObservedRunningTime="2025-12-04 14:19:12.027288117 +0000 UTC m=+1309.096006352" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.221613 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b79f64b59-wfklg"] Dec 04 14:19:12 crc kubenswrapper[4715]: E1204 14:19:12.222746 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb889380-79fe-461a-9a87-85c13fd8c716" containerName="keystone-bootstrap" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.222769 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb889380-79fe-461a-9a87-85c13fd8c716" containerName="keystone-bootstrap" Dec 04 14:19:12 crc kubenswrapper[4715]: E1204 14:19:12.222780 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="init" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.222786 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="init" Dec 04 14:19:12 crc kubenswrapper[4715]: E1204 14:19:12.222803 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="dnsmasq-dns" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.222813 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="dnsmasq-dns" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.223110 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb889380-79fe-461a-9a87-85c13fd8c716" containerName="keystone-bootstrap" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.223151 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="dnsmasq-dns" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.224402 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.233241 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.233434 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.233546 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.233717 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.233824 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.233972 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bskrn" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.252791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-internal-tls-certs\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.252865 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-fernet-keys\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.252957 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-public-tls-certs\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.252996 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-scripts\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.253022 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-credential-keys\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.253073 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-combined-ca-bundle\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.253182 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7p62\" (UniqueName: \"kubernetes.io/projected/a412f2fc-2452-41bc-9670-d0db2350f145-kube-api-access-s7p62\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.253225 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-config-data\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.253397 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b79f64b59-wfklg"] Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357199 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7p62\" (UniqueName: \"kubernetes.io/projected/a412f2fc-2452-41bc-9670-d0db2350f145-kube-api-access-s7p62\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-config-data\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-internal-tls-certs\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-fernet-keys\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357387 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-public-tls-certs\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-scripts\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357434 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-credential-keys\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.357457 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-combined-ca-bundle\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.364115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-config-data\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.364707 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-fernet-keys\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.366831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-combined-ca-bundle\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.370519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-public-tls-certs\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.378547 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-scripts\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.394133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-internal-tls-certs\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.395190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a412f2fc-2452-41bc-9670-d0db2350f145-credential-keys\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.401894 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7p62\" (UniqueName: \"kubernetes.io/projected/a412f2fc-2452-41bc-9670-d0db2350f145-kube-api-access-s7p62\") pod \"keystone-5b79f64b59-wfklg\" (UID: \"a412f2fc-2452-41bc-9670-d0db2350f145\") " pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:12 crc kubenswrapper[4715]: I1204 14:19:12.576760 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.168008 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.168200 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.214599 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" path="/var/lib/kubelet/pods/332031a0-392b-4497-9c04-9c1bbd0a2ea0/volumes" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.215534 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.215574 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.215918 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.252076 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.261351 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.263064 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.976094 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.976459 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.976476 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:19:13 crc kubenswrapper[4715]: I1204 14:19:13.976486 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:14 crc kubenswrapper[4715]: I1204 14:19:14.579398 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tmg6v" podUID="332031a0-392b-4497-9c04-9c1bbd0a2ea0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 04 14:19:14 crc kubenswrapper[4715]: I1204 14:19:14.743293 4715 scope.go:117] "RemoveContainer" containerID="e47a95d9a58803cf79875408ee672b661f48a6df5064e2c0c6065d030cfcfbda" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.113470 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8bbsj" event={"ID":"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02","Type":"ContainerDied","Data":"afe1eab2479b984b970b6b768e650c0d7366d5cff31df88c45304bd61310d3ba"} Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.113517 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe1eab2479b984b970b6b768e650c0d7366d5cff31df88c45304bd61310d3ba" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.153925 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.225683 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-db-sync-config-data\") pod \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.225734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-combined-ca-bundle\") pod \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.225807 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8g8g\" (UniqueName: \"kubernetes.io/projected/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-kube-api-access-g8g8g\") pod \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\" (UID: \"2ddb5944-20e2-4f19-96c4-ee93ac8cfa02\") " Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.236347 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-kube-api-access-g8g8g" (OuterVolumeSpecName: "kube-api-access-g8g8g") pod "2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" (UID: "2ddb5944-20e2-4f19-96c4-ee93ac8cfa02"). InnerVolumeSpecName "kube-api-access-g8g8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.236658 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" (UID: "2ddb5944-20e2-4f19-96c4-ee93ac8cfa02"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.307234 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" (UID: "2ddb5944-20e2-4f19-96c4-ee93ac8cfa02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.333104 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.333145 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.333158 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8g8g\" (UniqueName: \"kubernetes.io/projected/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02-kube-api-access-g8g8g\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:15 crc kubenswrapper[4715]: I1204 14:19:15.482425 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b79f64b59-wfklg"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.129083 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b79f64b59-wfklg" event={"ID":"a412f2fc-2452-41bc-9670-d0db2350f145","Type":"ContainerStarted","Data":"17cbb1c9bd89b74d6da1429d38724d4ec831df71fc6eed52d759aff8ffc7d073"} Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.129397 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b79f64b59-wfklg" event={"ID":"a412f2fc-2452-41bc-9670-d0db2350f145","Type":"ContainerStarted","Data":"b7eac9a36571d1772eba6955db3bdbe175ab459b69c1d3a4ffa460671d4f4d0f"} Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.130596 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.132994 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerStarted","Data":"fda30309ef0e04fbb9e89c36dd6a42cefd5ebbde0e791d60f1a6d2d7b742a8d3"} Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.136127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cgdwm" event={"ID":"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0","Type":"ContainerStarted","Data":"96ac05113e2494087f146b6b35619bf4ae4f1f9b50c7c629c76b100a8a2c3180"} Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.141652 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8bbsj" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.144332 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s5j8r" event={"ID":"7e311957-d00b-43da-a9dc-ca48386dfa88","Type":"ContainerStarted","Data":"2be8db361a9206eb161ca65372c60a8afb28cd4bd530b81bae4438f3887cf052"} Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.174443 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5b79f64b59-wfklg" podStartSLOduration=4.174425939 podStartE2EDuration="4.174425939s" podCreationTimestamp="2025-12-04 14:19:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:16.148729793 +0000 UTC m=+1313.217447998" watchObservedRunningTime="2025-12-04 14:19:16.174425939 +0000 UTC m=+1313.243144154" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.196429 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-cgdwm" podStartSLOduration=7.173039526 podStartE2EDuration="52.196405864s" podCreationTimestamp="2025-12-04 14:18:24 +0000 UTC" firstStartedPulling="2025-12-04 14:18:25.836918996 +0000 UTC m=+1262.905637211" lastFinishedPulling="2025-12-04 14:19:10.860285334 +0000 UTC m=+1307.929003549" observedRunningTime="2025-12-04 14:19:16.169461264 +0000 UTC m=+1313.238179479" watchObservedRunningTime="2025-12-04 14:19:16.196405864 +0000 UTC m=+1313.265124089" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.200872 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-s5j8r" podStartSLOduration=5.268222221 podStartE2EDuration="52.200855274s" podCreationTimestamp="2025-12-04 14:18:24 +0000 UTC" firstStartedPulling="2025-12-04 14:18:28.066930589 +0000 UTC m=+1265.135648804" lastFinishedPulling="2025-12-04 14:19:14.999563642 +0000 UTC m=+1312.068281857" observedRunningTime="2025-12-04 14:19:16.19551506 +0000 UTC m=+1313.264233285" watchObservedRunningTime="2025-12-04 14:19:16.200855274 +0000 UTC m=+1313.269573489" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.404105 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-64d967d55f-xm4d6"] Dec 04 14:19:16 crc kubenswrapper[4715]: E1204 14:19:16.404651 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" containerName="barbican-db-sync" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.404675 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" containerName="barbican-db-sync" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.404930 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" containerName="barbican-db-sync" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.406218 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.415080 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.415374 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-j8psz" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.418018 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.436823 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-f7449dddd-lkzvw"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.442856 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.448646 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.455788 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-combined-ca-bundle\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.455854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-config-data-custom\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.455920 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnqvt\" (UniqueName: \"kubernetes.io/projected/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-kube-api-access-pnqvt\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.455945 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-logs\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.456074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-config-data\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.468109 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-f7449dddd-lkzvw"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.481129 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64d967d55f-xm4d6"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.555867 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-j6rz7"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-combined-ca-bundle\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-config-data-custom\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557596 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnqvt\" (UniqueName: \"kubernetes.io/projected/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-kube-api-access-pnqvt\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-logs\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557801 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-config-data-custom\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-combined-ca-bundle\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557987 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589e9c2f-67f8-4208-aaa3-b391d59b7d01-logs\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.558124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzcjh\" (UniqueName: \"kubernetes.io/projected/589e9c2f-67f8-4208-aaa3-b391d59b7d01-kube-api-access-kzcjh\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.558237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-config-data\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.558349 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-config-data\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.557618 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.560108 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-logs\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.568519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-combined-ca-bundle\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.570029 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-config-data\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.583835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-config-data-custom\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.597056 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-j6rz7"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.597112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnqvt\" (UniqueName: \"kubernetes.io/projected/5a34d667-48aa-4bb9-82fa-bb528e9bcc20-kube-api-access-pnqvt\") pod \"barbican-worker-64d967d55f-xm4d6\" (UID: \"5a34d667-48aa-4bb9-82fa-bb528e9bcc20\") " pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.661773 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzcjh\" (UniqueName: \"kubernetes.io/projected/589e9c2f-67f8-4208-aaa3-b391d59b7d01-kube-api-access-kzcjh\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.661855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.661914 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.661994 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-config-data\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k52s\" (UniqueName: \"kubernetes.io/projected/92a3bae3-f661-492c-bb10-cf636c135d06-kube-api-access-7k52s\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662210 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-config\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662301 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-config-data-custom\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662332 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-combined-ca-bundle\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662406 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589e9c2f-67f8-4208-aaa3-b391d59b7d01-logs\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.662488 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.672434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589e9c2f-67f8-4208-aaa3-b391d59b7d01-logs\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.674975 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7896b66d76-jtzqq"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.675437 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-config-data-custom\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.699523 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-config-data\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.699712 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.717629 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.736087 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589e9c2f-67f8-4208-aaa3-b391d59b7d01-combined-ca-bundle\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.748448 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7896b66d76-jtzqq"] Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.749468 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzcjh\" (UniqueName: \"kubernetes.io/projected/589e9c2f-67f8-4208-aaa3-b391d59b7d01-kube-api-access-kzcjh\") pod \"barbican-keystone-listener-f7449dddd-lkzvw\" (UID: \"589e9c2f-67f8-4208-aaa3-b391d59b7d01\") " pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.765843 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2773b750-4db6-46e8-9337-4fd592cc8410-logs\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.765938 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k52s\" (UniqueName: \"kubernetes.io/projected/92a3bae3-f661-492c-bb10-cf636c135d06-kube-api-access-7k52s\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.765992 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data-custom\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766072 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-config\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766261 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9s95\" (UniqueName: \"kubernetes.io/projected/2773b750-4db6-46e8-9337-4fd592cc8410-kube-api-access-x9s95\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766371 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766427 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-combined-ca-bundle\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766481 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766509 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.766535 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.767856 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.778350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.779239 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-config\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.779845 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64d967d55f-xm4d6" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.780815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.781246 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.808120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.815513 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k52s\" (UniqueName: \"kubernetes.io/projected/92a3bae3-f661-492c-bb10-cf636c135d06-kube-api-access-7k52s\") pod \"dnsmasq-dns-586bdc5f9-j6rz7\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.869210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9s95\" (UniqueName: \"kubernetes.io/projected/2773b750-4db6-46e8-9337-4fd592cc8410-kube-api-access-x9s95\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.869333 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-combined-ca-bundle\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.869364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.869399 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2773b750-4db6-46e8-9337-4fd592cc8410-logs\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.869435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data-custom\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.875423 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2773b750-4db6-46e8-9337-4fd592cc8410-logs\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.875423 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-combined-ca-bundle\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.878851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.882743 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data-custom\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:16 crc kubenswrapper[4715]: I1204 14:19:16.889488 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9s95\" (UniqueName: \"kubernetes.io/projected/2773b750-4db6-46e8-9337-4fd592cc8410-kube-api-access-x9s95\") pod \"barbican-api-7896b66d76-jtzqq\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.002394 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.026690 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.400544 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.541458 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-f7449dddd-lkzvw"] Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.543938 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.558726 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64d967d55f-xm4d6"] Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.597248 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-j6rz7"] Dec 04 14:19:17 crc kubenswrapper[4715]: I1204 14:19:17.956323 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7896b66d76-jtzqq"] Dec 04 14:19:18 crc kubenswrapper[4715]: I1204 14:19:18.213952 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" event={"ID":"92a3bae3-f661-492c-bb10-cf636c135d06","Type":"ContainerStarted","Data":"067e8a0348f2a1d1fdf741f1bf0c2edf5a0b3ce12c0913a72a413bbc3db9554f"} Dec 04 14:19:18 crc kubenswrapper[4715]: I1204 14:19:18.270827 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64d967d55f-xm4d6" event={"ID":"5a34d667-48aa-4bb9-82fa-bb528e9bcc20","Type":"ContainerStarted","Data":"955042c66b7c060bb695804ddb5d65963670711edef21ab16b16f26f3c5592be"} Dec 04 14:19:18 crc kubenswrapper[4715]: I1204 14:19:18.294257 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7896b66d76-jtzqq" event={"ID":"2773b750-4db6-46e8-9337-4fd592cc8410","Type":"ContainerStarted","Data":"ef9376e0b8cade422a16d95bfb469628f9a2f44e7099cd517d398627003e7e58"} Dec 04 14:19:18 crc kubenswrapper[4715]: I1204 14:19:18.336060 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" event={"ID":"589e9c2f-67f8-4208-aaa3-b391d59b7d01","Type":"ContainerStarted","Data":"3424bcdaa1aecbbe555a2daaa9277615b0baa732cacf4a2e3684f4981fc11ec2"} Dec 04 14:19:18 crc kubenswrapper[4715]: I1204 14:19:18.772964 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 14:19:18 crc kubenswrapper[4715]: I1204 14:19:18.888249 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b798c69c6-g6fvm" podUID="8552f780-12e2-4eeb-bbc5-9ef637e9278f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.355738 4715 generic.go:334] "Generic (PLEG): container finished" podID="92a3bae3-f661-492c-bb10-cf636c135d06" containerID="336e9d52fb70bc3782a98e38285ecf20041d9a92a65f76c6640e39d24dc8d24c" exitCode=0 Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.355794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" event={"ID":"92a3bae3-f661-492c-bb10-cf636c135d06","Type":"ContainerDied","Data":"336e9d52fb70bc3782a98e38285ecf20041d9a92a65f76c6640e39d24dc8d24c"} Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.360874 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7896b66d76-jtzqq" event={"ID":"2773b750-4db6-46e8-9337-4fd592cc8410","Type":"ContainerStarted","Data":"18db19ae897873d8198e1589cd364976097faa3820993571a28b5385fc8b11d6"} Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.970741 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65b4dd5988-fzzwk"] Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.976147 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.983589 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 04 14:19:19 crc kubenswrapper[4715]: I1204 14:19:19.983818 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.014824 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65b4dd5988-fzzwk"] Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107067 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-config-data\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107191 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv4wq\" (UniqueName: \"kubernetes.io/projected/795cce68-d6db-4923-ab86-c474d5085318-kube-api-access-cv4wq\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107252 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/795cce68-d6db-4923-ab86-c474d5085318-logs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107350 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-combined-ca-bundle\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107453 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-config-data-custom\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107555 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-internal-tls-certs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.107587 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-public-tls-certs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.226364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-combined-ca-bundle\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.226456 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-config-data-custom\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.226529 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-internal-tls-certs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.226554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-public-tls-certs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.226643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-config-data\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.228533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv4wq\" (UniqueName: \"kubernetes.io/projected/795cce68-d6db-4923-ab86-c474d5085318-kube-api-access-cv4wq\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.228599 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/795cce68-d6db-4923-ab86-c474d5085318-logs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.229370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/795cce68-d6db-4923-ab86-c474d5085318-logs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.238164 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-combined-ca-bundle\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.238397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-config-data-custom\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.263895 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-internal-tls-certs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.271651 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv4wq\" (UniqueName: \"kubernetes.io/projected/795cce68-d6db-4923-ab86-c474d5085318-kube-api-access-cv4wq\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.278908 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-config-data\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.301012 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/795cce68-d6db-4923-ab86-c474d5085318-public-tls-certs\") pod \"barbican-api-65b4dd5988-fzzwk\" (UID: \"795cce68-d6db-4923-ab86-c474d5085318\") " pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.318676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.401231 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7896b66d76-jtzqq" event={"ID":"2773b750-4db6-46e8-9337-4fd592cc8410","Type":"ContainerStarted","Data":"31b336103582571f529a66fd90896aefebf375cbf786c57d0089b2fe75d25b45"} Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.401976 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.402059 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.419158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" event={"ID":"92a3bae3-f661-492c-bb10-cf636c135d06","Type":"ContainerStarted","Data":"4249d77a08662aebb6db84ab2ea51b79e5a7ea600284207d7c0c9160afb94728"} Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.419952 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.448168 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7896b66d76-jtzqq" podStartSLOduration=4.448143748 podStartE2EDuration="4.448143748s" podCreationTimestamp="2025-12-04 14:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:20.437738676 +0000 UTC m=+1317.506456891" watchObservedRunningTime="2025-12-04 14:19:20.448143748 +0000 UTC m=+1317.516861963" Dec 04 14:19:20 crc kubenswrapper[4715]: I1204 14:19:20.479949 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" podStartSLOduration=4.479927029 podStartE2EDuration="4.479927029s" podCreationTimestamp="2025-12-04 14:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:20.476493486 +0000 UTC m=+1317.545211711" watchObservedRunningTime="2025-12-04 14:19:20.479927029 +0000 UTC m=+1317.548645254" Dec 04 14:19:21 crc kubenswrapper[4715]: I1204 14:19:21.004477 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:19:21 crc kubenswrapper[4715]: I1204 14:19:21.441771 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:19:22 crc kubenswrapper[4715]: I1204 14:19:22.445263 4715 generic.go:334] "Generic (PLEG): container finished" podID="7e311957-d00b-43da-a9dc-ca48386dfa88" containerID="2be8db361a9206eb161ca65372c60a8afb28cd4bd530b81bae4438f3887cf052" exitCode=0 Dec 04 14:19:22 crc kubenswrapper[4715]: I1204 14:19:22.445402 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s5j8r" event={"ID":"7e311957-d00b-43da-a9dc-ca48386dfa88","Type":"ContainerDied","Data":"2be8db361a9206eb161ca65372c60a8afb28cd4bd530b81bae4438f3887cf052"} Dec 04 14:19:22 crc kubenswrapper[4715]: I1204 14:19:22.984663 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65b4dd5988-fzzwk"] Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.457207 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b4dd5988-fzzwk" event={"ID":"795cce68-d6db-4923-ab86-c474d5085318","Type":"ContainerStarted","Data":"4f9ade4eebaaad257f5d2e28abd2d1e52c964f46a7f9449b5a25b7ac0cf15258"} Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.457904 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b4dd5988-fzzwk" event={"ID":"795cce68-d6db-4923-ab86-c474d5085318","Type":"ContainerStarted","Data":"3b944980f741de98af77290a4d94f557436095022c38d2a13cb62dcec846cfb9"} Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.482150 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64d967d55f-xm4d6" event={"ID":"5a34d667-48aa-4bb9-82fa-bb528e9bcc20","Type":"ContainerStarted","Data":"ef0dfb9ab17b4f8b87368287991341ca32cb20c0ae9ef3dd5c3c62b417e8523a"} Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.482412 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64d967d55f-xm4d6" event={"ID":"5a34d667-48aa-4bb9-82fa-bb528e9bcc20","Type":"ContainerStarted","Data":"b275db6ac9ac9b0a6c5ed5d63212c346c13ed553ec0e652ffb664346ecfd7f9f"} Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.502862 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" event={"ID":"589e9c2f-67f8-4208-aaa3-b391d59b7d01","Type":"ContainerStarted","Data":"1ca57de0f9565b72d1f6a9bc7a83f6da78b12e7769fee93887325c0dc31e8575"} Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.503015 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" event={"ID":"589e9c2f-67f8-4208-aaa3-b391d59b7d01","Type":"ContainerStarted","Data":"2c54fa6ef85992473757f967d616a0b3330dd4f6f115d173f3639047643dbc12"} Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.521095 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-64d967d55f-xm4d6" podStartSLOduration=2.563778237 podStartE2EDuration="7.521073298s" podCreationTimestamp="2025-12-04 14:19:16 +0000 UTC" firstStartedPulling="2025-12-04 14:19:17.602193685 +0000 UTC m=+1314.670911900" lastFinishedPulling="2025-12-04 14:19:22.559488736 +0000 UTC m=+1319.628206961" observedRunningTime="2025-12-04 14:19:23.517494801 +0000 UTC m=+1320.586213006" watchObservedRunningTime="2025-12-04 14:19:23.521073298 +0000 UTC m=+1320.589791523" Dec 04 14:19:23 crc kubenswrapper[4715]: I1204 14:19:23.556791 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-f7449dddd-lkzvw" podStartSLOduration=2.599602157 podStartE2EDuration="7.556765814s" podCreationTimestamp="2025-12-04 14:19:16 +0000 UTC" firstStartedPulling="2025-12-04 14:19:17.604660712 +0000 UTC m=+1314.673378927" lastFinishedPulling="2025-12-04 14:19:22.561824369 +0000 UTC m=+1319.630542584" observedRunningTime="2025-12-04 14:19:23.552977112 +0000 UTC m=+1320.621695347" watchObservedRunningTime="2025-12-04 14:19:23.556765814 +0000 UTC m=+1320.625484049" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.039615 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s5j8r" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.130584 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-combined-ca-bundle\") pod \"7e311957-d00b-43da-a9dc-ca48386dfa88\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.131133 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh7fj\" (UniqueName: \"kubernetes.io/projected/7e311957-d00b-43da-a9dc-ca48386dfa88-kube-api-access-nh7fj\") pod \"7e311957-d00b-43da-a9dc-ca48386dfa88\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.131316 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e311957-d00b-43da-a9dc-ca48386dfa88-logs\") pod \"7e311957-d00b-43da-a9dc-ca48386dfa88\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.131460 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-scripts\") pod \"7e311957-d00b-43da-a9dc-ca48386dfa88\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.131629 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-config-data\") pod \"7e311957-d00b-43da-a9dc-ca48386dfa88\" (UID: \"7e311957-d00b-43da-a9dc-ca48386dfa88\") " Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.137998 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e311957-d00b-43da-a9dc-ca48386dfa88-logs" (OuterVolumeSpecName: "logs") pod "7e311957-d00b-43da-a9dc-ca48386dfa88" (UID: "7e311957-d00b-43da-a9dc-ca48386dfa88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.148260 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e311957-d00b-43da-a9dc-ca48386dfa88-kube-api-access-nh7fj" (OuterVolumeSpecName: "kube-api-access-nh7fj") pod "7e311957-d00b-43da-a9dc-ca48386dfa88" (UID: "7e311957-d00b-43da-a9dc-ca48386dfa88"). InnerVolumeSpecName "kube-api-access-nh7fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.163297 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-scripts" (OuterVolumeSpecName: "scripts") pod "7e311957-d00b-43da-a9dc-ca48386dfa88" (UID: "7e311957-d00b-43da-a9dc-ca48386dfa88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.186727 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-config-data" (OuterVolumeSpecName: "config-data") pod "7e311957-d00b-43da-a9dc-ca48386dfa88" (UID: "7e311957-d00b-43da-a9dc-ca48386dfa88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.188126 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e311957-d00b-43da-a9dc-ca48386dfa88" (UID: "7e311957-d00b-43da-a9dc-ca48386dfa88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.236387 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh7fj\" (UniqueName: \"kubernetes.io/projected/7e311957-d00b-43da-a9dc-ca48386dfa88-kube-api-access-nh7fj\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.236432 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e311957-d00b-43da-a9dc-ca48386dfa88-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.236445 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.236456 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.236468 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e311957-d00b-43da-a9dc-ca48386dfa88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.526861 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s5j8r" event={"ID":"7e311957-d00b-43da-a9dc-ca48386dfa88","Type":"ContainerDied","Data":"b2955330a8ff7a6416467c50f8c921b8d8ce53ee79c34adf4286d969d28a08ea"} Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.526903 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2955330a8ff7a6416467c50f8c921b8d8ce53ee79c34adf4286d969d28a08ea" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.526970 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s5j8r" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.551609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b4dd5988-fzzwk" event={"ID":"795cce68-d6db-4923-ab86-c474d5085318","Type":"ContainerStarted","Data":"22a6b4445adf14e2c4717b6c5237001519e3569a53f3aadc43e6b60281ce83fa"} Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.552223 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.553101 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.586134 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65f6dc6d96-gj9d4"] Dec 04 14:19:24 crc kubenswrapper[4715]: E1204 14:19:24.586658 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e311957-d00b-43da-a9dc-ca48386dfa88" containerName="placement-db-sync" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.586677 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e311957-d00b-43da-a9dc-ca48386dfa88" containerName="placement-db-sync" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.586929 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e311957-d00b-43da-a9dc-ca48386dfa88" containerName="placement-db-sync" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.588156 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.595854 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65b4dd5988-fzzwk" podStartSLOduration=5.595826534 podStartE2EDuration="5.595826534s" podCreationTimestamp="2025-12-04 14:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:24.580413246 +0000 UTC m=+1321.649131471" watchObservedRunningTime="2025-12-04 14:19:24.595826534 +0000 UTC m=+1321.664544759" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.596767 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.597104 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.599382 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qljnv" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.599769 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.610027 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.689240 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65f6dc6d96-gj9d4"] Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.757958 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-public-tls-certs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.758008 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-internal-tls-certs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.758083 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad8392c-fc50-4cd5-af7c-9d186021d336-logs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.758153 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-scripts\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.758242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-combined-ca-bundle\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.758264 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2594\" (UniqueName: \"kubernetes.io/projected/aad8392c-fc50-4cd5-af7c-9d186021d336-kube-api-access-x2594\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.758293 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-config-data\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.859683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad8392c-fc50-4cd5-af7c-9d186021d336-logs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.859805 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-scripts\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.859865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-combined-ca-bundle\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.859894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2594\" (UniqueName: \"kubernetes.io/projected/aad8392c-fc50-4cd5-af7c-9d186021d336-kube-api-access-x2594\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.859929 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-config-data\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.860023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-public-tls-certs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.860191 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-internal-tls-certs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.860305 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad8392c-fc50-4cd5-af7c-9d186021d336-logs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.865978 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-combined-ca-bundle\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.870928 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-scripts\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.871462 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-config-data\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.873685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-internal-tls-certs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.879461 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aad8392c-fc50-4cd5-af7c-9d186021d336-public-tls-certs\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.890678 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2594\" (UniqueName: \"kubernetes.io/projected/aad8392c-fc50-4cd5-af7c-9d186021d336-kube-api-access-x2594\") pod \"placement-65f6dc6d96-gj9d4\" (UID: \"aad8392c-fc50-4cd5-af7c-9d186021d336\") " pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:24 crc kubenswrapper[4715]: I1204 14:19:24.924263 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:25 crc kubenswrapper[4715]: I1204 14:19:25.461600 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65f6dc6d96-gj9d4"] Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.005523 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.067937 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z4zqf"] Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.068178 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="dnsmasq-dns" containerID="cri-o://bdd78e6d8ce2e391a2a3c679b069dc2b9e4dc37f274bf1dbe8a52828ada6e4b2" gracePeriod=10 Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.595815 4715 generic.go:334] "Generic (PLEG): container finished" podID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" containerID="96ac05113e2494087f146b6b35619bf4ae4f1f9b50c7c629c76b100a8a2c3180" exitCode=0 Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.595893 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cgdwm" event={"ID":"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0","Type":"ContainerDied","Data":"96ac05113e2494087f146b6b35619bf4ae4f1f9b50c7c629c76b100a8a2c3180"} Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.598524 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerID="bdd78e6d8ce2e391a2a3c679b069dc2b9e4dc37f274bf1dbe8a52828ada6e4b2" exitCode=0 Dec 04 14:19:27 crc kubenswrapper[4715]: I1204 14:19:27.598551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" event={"ID":"4e56f804-b6ae-4d86-8e0b-9aed08a299b4","Type":"ContainerDied","Data":"bdd78e6d8ce2e391a2a3c679b069dc2b9e4dc37f274bf1dbe8a52828ada6e4b2"} Dec 04 14:19:28 crc kubenswrapper[4715]: I1204 14:19:28.615483 4715 generic.go:334] "Generic (PLEG): container finished" podID="95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" containerID="ceb9e9fdc0fbe0e01c1cff1e55fde32653535576aeb494e4e54de204faf484cd" exitCode=0 Dec 04 14:19:28 crc kubenswrapper[4715]: I1204 14:19:28.615569 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5mrnk" event={"ID":"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1","Type":"ContainerDied","Data":"ceb9e9fdc0fbe0e01c1cff1e55fde32653535576aeb494e4e54de204faf484cd"} Dec 04 14:19:28 crc kubenswrapper[4715]: I1204 14:19:28.771812 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 14:19:28 crc kubenswrapper[4715]: I1204 14:19:28.884468 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b798c69c6-g6fvm" podUID="8552f780-12e2-4eeb-bbc5-9ef637e9278f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 04 14:19:29 crc kubenswrapper[4715]: I1204 14:19:29.148158 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:29 crc kubenswrapper[4715]: I1204 14:19:29.486701 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:29 crc kubenswrapper[4715]: I1204 14:19:29.728411 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.651388 4715 generic.go:334] "Generic (PLEG): container finished" podID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerID="207bb21cc12f37dde2b5a2635bd91b784400406326add9b1bbfa01c6fe3a1eb4" exitCode=137 Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.651883 4715 generic.go:334] "Generic (PLEG): container finished" podID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerID="dbb5250430bbee33218c769d14a9d9d05e1495c461ecb329525091ccaaed8b15" exitCode=137 Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.651474 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8599dd67c5-qs56p" event={"ID":"e8331eb7-653b-4a6c-ada8-f13dcf916a8a","Type":"ContainerDied","Data":"207bb21cc12f37dde2b5a2635bd91b784400406326add9b1bbfa01c6fe3a1eb4"} Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.651962 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8599dd67c5-qs56p" event={"ID":"e8331eb7-653b-4a6c-ada8-f13dcf916a8a","Type":"ContainerDied","Data":"dbb5250430bbee33218c769d14a9d9d05e1495c461ecb329525091ccaaed8b15"} Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.653974 4715 generic.go:334] "Generic (PLEG): container finished" podID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerID="cde5fc09e99af07fca1b14c1394b5942eed07a676cad8a6acb20ea77d12058e6" exitCode=137 Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.653993 4715 generic.go:334] "Generic (PLEG): container finished" podID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerID="6c96e1ac3cfcae2163fd4c6676397b563df5edeb5e4005f1c1459da15c59f831" exitCode=137 Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.654027 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57458fcb8f-w827q" event={"ID":"a1c0859d-f9ae-457f-9c44-ea9891d7756b","Type":"ContainerDied","Data":"cde5fc09e99af07fca1b14c1394b5942eed07a676cad8a6acb20ea77d12058e6"} Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.654074 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57458fcb8f-w827q" event={"ID":"a1c0859d-f9ae-457f-9c44-ea9891d7756b","Type":"ContainerDied","Data":"6c96e1ac3cfcae2163fd4c6676397b563df5edeb5e4005f1c1459da15c59f831"} Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.655866 4715 generic.go:334] "Generic (PLEG): container finished" podID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerID="d9d5504dd22df7cd3fddb8b08067bef9d4f4ee6732924b53b10b4c34fa598d9e" exitCode=137 Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.655900 4715 generic.go:334] "Generic (PLEG): container finished" podID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerID="80bb690535d08957f268fb451a9b2608f1e5529fbe003046329fd5dbf9c7373b" exitCode=137 Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.655924 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5978d7d969-7vwfh" event={"ID":"0a909008-f754-49fb-87cb-dc0d09c91bfd","Type":"ContainerDied","Data":"d9d5504dd22df7cd3fddb8b08067bef9d4f4ee6732924b53b10b4c34fa598d9e"} Dec 04 14:19:31 crc kubenswrapper[4715]: I1204 14:19:31.655956 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5978d7d969-7vwfh" event={"ID":"0a909008-f754-49fb-87cb-dc0d09c91bfd","Type":"ContainerDied","Data":"80bb690535d08957f268fb451a9b2608f1e5529fbe003046329fd5dbf9c7373b"} Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.210132 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.330598 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.366235 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.381422 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65b4dd5988-fzzwk" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.444753 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-combined-ca-bundle\") pod \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.444896 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxdvg\" (UniqueName: \"kubernetes.io/projected/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-kube-api-access-fxdvg\") pod \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.445092 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-config\") pod \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\" (UID: \"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.470311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-kube-api-access-fxdvg" (OuterVolumeSpecName: "kube-api-access-fxdvg") pod "95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" (UID: "95ea9b8a-207e-4e7b-9df2-c51ccc497fa1"). InnerVolumeSpecName "kube-api-access-fxdvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.482935 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7896b66d76-jtzqq"] Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.483620 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7896b66d76-jtzqq" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api-log" containerID="cri-o://18db19ae897873d8198e1589cd364976097faa3820993571a28b5385fc8b11d6" gracePeriod=30 Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.483723 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7896b66d76-jtzqq" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api" containerID="cri-o://31b336103582571f529a66fd90896aefebf375cbf786c57d0089b2fe75d25b45" gracePeriod=30 Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.492382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-config" (OuterVolumeSpecName: "config") pod "95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" (UID: "95ea9b8a-207e-4e7b-9df2-c51ccc497fa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.521469 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" (UID: "95ea9b8a-207e-4e7b-9df2-c51ccc497fa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546373 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-config-data\") pod \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546446 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-db-sync-config-data\") pod \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546538 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-etc-machine-id\") pod \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546622 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-combined-ca-bundle\") pod \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546725 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jqbz\" (UniqueName: \"kubernetes.io/projected/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-kube-api-access-2jqbz\") pod \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-scripts\") pod \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\" (UID: \"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0\") " Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.546958 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" (UID: "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.547556 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.547591 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.547603 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.547614 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxdvg\" (UniqueName: \"kubernetes.io/projected/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1-kube-api-access-fxdvg\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.558438 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" (UID: "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.560272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-kube-api-access-2jqbz" (OuterVolumeSpecName: "kube-api-access-2jqbz") pod "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" (UID: "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0"). InnerVolumeSpecName "kube-api-access-2jqbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.561330 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-scripts" (OuterVolumeSpecName: "scripts") pod "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" (UID: "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.607275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" (UID: "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.652317 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.652373 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jqbz\" (UniqueName: \"kubernetes.io/projected/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-kube-api-access-2jqbz\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.652390 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.652402 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.666634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-config-data" (OuterVolumeSpecName: "config-data") pod "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" (UID: "2ca3112c-c14e-41af-ad25-c16c7ca1a8c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.694396 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6dc6d96-gj9d4" event={"ID":"aad8392c-fc50-4cd5-af7c-9d186021d336","Type":"ContainerStarted","Data":"bb6c1caff514b72e36f5e3f3ab7d766025cd73180882231a33556ab7494050db"} Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.716515 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5mrnk" event={"ID":"95ea9b8a-207e-4e7b-9df2-c51ccc497fa1","Type":"ContainerDied","Data":"db414aa09f958319c2529dfb137c1ccf39473c769761e83052843c41c224aec9"} Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.716567 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db414aa09f958319c2529dfb137c1ccf39473c769761e83052843c41c224aec9" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.716644 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5mrnk" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.721850 4715 generic.go:334] "Generic (PLEG): container finished" podID="2773b750-4db6-46e8-9337-4fd592cc8410" containerID="18db19ae897873d8198e1589cd364976097faa3820993571a28b5385fc8b11d6" exitCode=143 Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.721917 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7896b66d76-jtzqq" event={"ID":"2773b750-4db6-46e8-9337-4fd592cc8410","Type":"ContainerDied","Data":"18db19ae897873d8198e1589cd364976097faa3820993571a28b5385fc8b11d6"} Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.724824 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cgdwm" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.724898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cgdwm" event={"ID":"2ca3112c-c14e-41af-ad25-c16c7ca1a8c0","Type":"ContainerDied","Data":"47a3005461d3c23508fcdfc14b3c9c8582ca422ee1c6ba0242d407d0f374aa3b"} Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.724933 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47a3005461d3c23508fcdfc14b3c9c8582ca422ee1c6ba0242d407d0f374aa3b" Dec 04 14:19:32 crc kubenswrapper[4715]: I1204 14:19:32.753463 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.538615 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-bcn2w"] Dec 04 14:19:33 crc kubenswrapper[4715]: E1204 14:19:33.539436 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" containerName="neutron-db-sync" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.539464 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" containerName="neutron-db-sync" Dec 04 14:19:33 crc kubenswrapper[4715]: E1204 14:19:33.539489 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" containerName="cinder-db-sync" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.539499 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" containerName="cinder-db-sync" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.539751 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" containerName="neutron-db-sync" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.539782 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" containerName="cinder-db-sync" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.541101 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.621586 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-bcn2w"] Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.684206 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.684283 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-svc\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.684320 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.684440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-config\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.684527 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.684608 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h2cb\" (UniqueName: \"kubernetes.io/projected/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-kube-api-access-5h2cb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.787210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-config\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.787356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.787471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h2cb\" (UniqueName: \"kubernetes.io/projected/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-kube-api-access-5h2cb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.787515 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.787578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-svc\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.787639 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.789406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-svc\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.789459 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-config\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.789481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.789501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.790060 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.829525 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.831395 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.852432 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.852681 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.852793 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.852905 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zvl22" Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.901670 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:33 crc kubenswrapper[4715]: I1204 14:19:33.915454 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h2cb\" (UniqueName: \"kubernetes.io/projected/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-kube-api-access-5h2cb\") pod \"dnsmasq-dns-85ff748b95-bcn2w\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:33.998503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:33.998558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:33.998593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:33.998670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:33.998750 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dznxb\" (UniqueName: \"kubernetes.io/projected/c630f1e1-b605-4282-b136-647528bc86b4-kube-api-access-dznxb\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:33.999118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c630f1e1-b605-4282-b136-647528bc86b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.053175 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d7bd77bfb-nlk9m"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.065205 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.082071 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.098742 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.101763 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c630f1e1-b605-4282-b136-647528bc86b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.101912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.101947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.101987 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.102105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.102210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dznxb\" (UniqueName: \"kubernetes.io/projected/c630f1e1-b605-4282-b136-647528bc86b4-kube-api-access-dznxb\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.106399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c630f1e1-b605-4282-b136-647528bc86b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.142989 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.143497 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ck2jj" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.152084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.157476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.171117 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.207777 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dznxb\" (UniqueName: \"kubernetes.io/projected/c630f1e1-b605-4282-b136-647528bc86b4-kube-api-access-dznxb\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.215951 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.217544 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-ovndb-tls-certs\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.219902 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlh4v\" (UniqueName: \"kubernetes.io/projected/d49eba49-c182-4ec6-9a61-d14cddf541fd-kube-api-access-dlh4v\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.220077 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-config\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.220209 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-httpd-config\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.220341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-combined-ca-bundle\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.232941 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.273152 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d7bd77bfb-nlk9m"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.321720 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.323280 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlh4v\" (UniqueName: \"kubernetes.io/projected/d49eba49-c182-4ec6-9a61-d14cddf541fd-kube-api-access-dlh4v\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.323347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-config\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.323372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-httpd-config\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.323405 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-combined-ca-bundle\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.323520 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-ovndb-tls-certs\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.336154 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-config\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.343519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-httpd-config\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.344250 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-ovndb-tls-certs\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.353301 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-combined-ca-bundle\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.375384 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-bcn2w"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.376364 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlh4v\" (UniqueName: \"kubernetes.io/projected/d49eba49-c182-4ec6-9a61-d14cddf541fd-kube-api-access-dlh4v\") pod \"neutron-d7bd77bfb-nlk9m\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.397092 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-swb86"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.398904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.445260 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-swb86"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.485163 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.487275 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.491419 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.501336 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.515214 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.529154 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-config\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.529216 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.529267 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4rtq\" (UniqueName: \"kubernetes.io/projected/38ade7f7-11d1-43a2-9520-f6b217314df9-kube-api-access-f4rtq\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.529361 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.529433 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.529498 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635206 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4rtq\" (UniqueName: \"kubernetes.io/projected/38ade7f7-11d1-43a2-9520-f6b217314df9-kube-api-access-f4rtq\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635403 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79b4bf01-31b0-4a13-9359-8574e4fba640-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635425 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-scripts\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635467 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b4bf01-31b0-4a13-9359-8574e4fba640-logs\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635524 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635576 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxd9s\" (UniqueName: \"kubernetes.io/projected/79b4bf01-31b0-4a13-9359-8574e4fba640-kube-api-access-vxd9s\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635648 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635745 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635782 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data-custom\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.635806 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-config\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.636957 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.637061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-config\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.637935 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.644159 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.646116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.661927 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4rtq\" (UniqueName: \"kubernetes.io/projected/38ade7f7-11d1-43a2-9520-f6b217314df9-kube-api-access-f4rtq\") pod \"dnsmasq-dns-5c9776ccc5-swb86\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737229 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737308 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737338 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data-custom\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737399 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79b4bf01-31b0-4a13-9359-8574e4fba640-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737428 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-scripts\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b4bf01-31b0-4a13-9359-8574e4fba640-logs\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.737503 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxd9s\" (UniqueName: \"kubernetes.io/projected/79b4bf01-31b0-4a13-9359-8574e4fba640-kube-api-access-vxd9s\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.742576 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.743889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79b4bf01-31b0-4a13-9359-8574e4fba640-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.746359 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b4bf01-31b0-4a13-9359-8574e4fba640-logs\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.748964 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-scripts\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.755004 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.760116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data-custom\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: E1204 14:19:34.765106 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Dec 04 14:19:34 crc kubenswrapper[4715]: E1204 14:19:34.765320 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqtrm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(8b94f46e-2d75-479c-9ef6-b8fac3332e2f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.765592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxd9s\" (UniqueName: \"kubernetes.io/projected/79b4bf01-31b0-4a13-9359-8574e4fba640-kube-api-access-vxd9s\") pod \"cinder-api-0\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: E1204 14:19:34.766430 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.769212 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.770675 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57458fcb8f-w827q" event={"ID":"a1c0859d-f9ae-457f-9c44-ea9891d7756b","Type":"ContainerDied","Data":"f951019dbe81cf726cb7dc69baa250b984a5ccc0eae9f3b7aa00edabc6ea9306"} Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.770733 4715 scope.go:117] "RemoveContainer" containerID="cde5fc09e99af07fca1b14c1394b5942eed07a676cad8a6acb20ea77d12058e6" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.779367 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5978d7d969-7vwfh" event={"ID":"0a909008-f754-49fb-87cb-dc0d09c91bfd","Type":"ContainerDied","Data":"6a6c7dbc250388b68fbbcc2528f953c90c29f88b244e2c257b35095b2da86c4d"} Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.779695 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a6c7dbc250388b68fbbcc2528f953c90c29f88b244e2c257b35095b2da86c4d" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.792784 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.799156 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" event={"ID":"4e56f804-b6ae-4d86-8e0b-9aed08a299b4","Type":"ContainerDied","Data":"88b8d4f78d7dc50dd59ee57325750e751996ee854cdcef8f39f0cf8fc3d1e6c8"} Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.801326 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.819584 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.819809 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8599dd67c5-qs56p" event={"ID":"e8331eb7-653b-4a6c-ada8-f13dcf916a8a","Type":"ContainerDied","Data":"ff111cd200947cb410cf03c26a74ef33d037230944e0df444cd3f0a71c55ca0c"} Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.838503 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm5x9\" (UniqueName: \"kubernetes.io/projected/a1c0859d-f9ae-457f-9c44-ea9891d7756b-kube-api-access-fm5x9\") pod \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.838604 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1c0859d-f9ae-457f-9c44-ea9891d7756b-horizon-secret-key\") pod \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.838652 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c0859d-f9ae-457f-9c44-ea9891d7756b-logs\") pod \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.838675 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-scripts\") pod \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.838692 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-config-data\") pod \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\" (UID: \"a1c0859d-f9ae-457f-9c44-ea9891d7756b\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.843567 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c0859d-f9ae-457f-9c44-ea9891d7756b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a1c0859d-f9ae-457f-9c44-ea9891d7756b" (UID: "a1c0859d-f9ae-457f-9c44-ea9891d7756b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.843872 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c0859d-f9ae-457f-9c44-ea9891d7756b-logs" (OuterVolumeSpecName: "logs") pod "a1c0859d-f9ae-457f-9c44-ea9891d7756b" (UID: "a1c0859d-f9ae-457f-9c44-ea9891d7756b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.848767 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.855254 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c0859d-f9ae-457f-9c44-ea9891d7756b-kube-api-access-fm5x9" (OuterVolumeSpecName: "kube-api-access-fm5x9") pod "a1c0859d-f9ae-457f-9c44-ea9891d7756b" (UID: "a1c0859d-f9ae-457f-9c44-ea9891d7756b"). InnerVolumeSpecName "kube-api-access-fm5x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.855766 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.897529 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-config-data" (OuterVolumeSpecName: "config-data") pod "a1c0859d-f9ae-457f-9c44-ea9891d7756b" (UID: "a1c0859d-f9ae-457f-9c44-ea9891d7756b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.938665 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-scripts" (OuterVolumeSpecName: "scripts") pod "a1c0859d-f9ae-457f-9c44-ea9891d7756b" (UID: "a1c0859d-f9ae-457f-9c44-ea9891d7756b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.941672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-swift-storage-0\") pod \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.941825 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a909008-f754-49fb-87cb-dc0d09c91bfd-logs\") pod \"0a909008-f754-49fb-87cb-dc0d09c91bfd\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.941915 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-config\") pod \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942012 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-logs\") pod \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942145 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-config-data\") pod \"0a909008-f754-49fb-87cb-dc0d09c91bfd\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-horizon-secret-key\") pod \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942308 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-sb\") pod \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942383 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhtgt\" (UniqueName: \"kubernetes.io/projected/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-kube-api-access-lhtgt\") pod \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942455 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-config-data\") pod \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942545 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-scripts\") pod \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942655 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp9nm\" (UniqueName: \"kubernetes.io/projected/0a909008-f754-49fb-87cb-dc0d09c91bfd-kube-api-access-bp9nm\") pod \"0a909008-f754-49fb-87cb-dc0d09c91bfd\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942720 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-nb\") pod \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-svc\") pod \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\" (UID: \"4e56f804-b6ae-4d86-8e0b-9aed08a299b4\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.942968 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2psbw\" (UniqueName: \"kubernetes.io/projected/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-kube-api-access-2psbw\") pod \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\" (UID: \"e8331eb7-653b-4a6c-ada8-f13dcf916a8a\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943057 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-scripts\") pod \"0a909008-f754-49fb-87cb-dc0d09c91bfd\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943151 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a909008-f754-49fb-87cb-dc0d09c91bfd-horizon-secret-key\") pod \"0a909008-f754-49fb-87cb-dc0d09c91bfd\" (UID: \"0a909008-f754-49fb-87cb-dc0d09c91bfd\") " Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943655 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm5x9\" (UniqueName: \"kubernetes.io/projected/a1c0859d-f9ae-457f-9c44-ea9891d7756b-kube-api-access-fm5x9\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943735 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1c0859d-f9ae-457f-9c44-ea9891d7756b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943790 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c0859d-f9ae-457f-9c44-ea9891d7756b-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943848 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.943902 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1c0859d-f9ae-457f-9c44-ea9891d7756b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.946449 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-logs" (OuterVolumeSpecName: "logs") pod "e8331eb7-653b-4a6c-ada8-f13dcf916a8a" (UID: "e8331eb7-653b-4a6c-ada8-f13dcf916a8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.957738 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a909008-f754-49fb-87cb-dc0d09c91bfd-logs" (OuterVolumeSpecName: "logs") pod "0a909008-f754-49fb-87cb-dc0d09c91bfd" (UID: "0a909008-f754-49fb-87cb-dc0d09c91bfd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.964937 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-kube-api-access-2psbw" (OuterVolumeSpecName: "kube-api-access-2psbw") pod "e8331eb7-653b-4a6c-ada8-f13dcf916a8a" (UID: "e8331eb7-653b-4a6c-ada8-f13dcf916a8a"). InnerVolumeSpecName "kube-api-access-2psbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.979294 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e8331eb7-653b-4a6c-ada8-f13dcf916a8a" (UID: "e8331eb7-653b-4a6c-ada8-f13dcf916a8a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.983343 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-kube-api-access-lhtgt" (OuterVolumeSpecName: "kube-api-access-lhtgt") pod "4e56f804-b6ae-4d86-8e0b-9aed08a299b4" (UID: "4e56f804-b6ae-4d86-8e0b-9aed08a299b4"). InnerVolumeSpecName "kube-api-access-lhtgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:34 crc kubenswrapper[4715]: I1204 14:19:34.991151 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a909008-f754-49fb-87cb-dc0d09c91bfd-kube-api-access-bp9nm" (OuterVolumeSpecName: "kube-api-access-bp9nm") pod "0a909008-f754-49fb-87cb-dc0d09c91bfd" (UID: "0a909008-f754-49fb-87cb-dc0d09c91bfd"). InnerVolumeSpecName "kube-api-access-bp9nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.002808 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a909008-f754-49fb-87cb-dc0d09c91bfd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0a909008-f754-49fb-87cb-dc0d09c91bfd" (UID: "0a909008-f754-49fb-87cb-dc0d09c91bfd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045487 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2psbw\" (UniqueName: \"kubernetes.io/projected/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-kube-api-access-2psbw\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045521 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a909008-f754-49fb-87cb-dc0d09c91bfd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045535 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a909008-f754-49fb-87cb-dc0d09c91bfd-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045548 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045559 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045570 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhtgt\" (UniqueName: \"kubernetes.io/projected/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-kube-api-access-lhtgt\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.045581 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp9nm\" (UniqueName: \"kubernetes.io/projected/0a909008-f754-49fb-87cb-dc0d09c91bfd-kube-api-access-bp9nm\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.061577 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-config-data" (OuterVolumeSpecName: "config-data") pod "0a909008-f754-49fb-87cb-dc0d09c91bfd" (UID: "0a909008-f754-49fb-87cb-dc0d09c91bfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.117394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-scripts" (OuterVolumeSpecName: "scripts") pod "0a909008-f754-49fb-87cb-dc0d09c91bfd" (UID: "0a909008-f754-49fb-87cb-dc0d09c91bfd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.126723 4715 scope.go:117] "RemoveContainer" containerID="6c96e1ac3cfcae2163fd4c6676397b563df5edeb5e4005f1c1459da15c59f831" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.133536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4e56f804-b6ae-4d86-8e0b-9aed08a299b4" (UID: "4e56f804-b6ae-4d86-8e0b-9aed08a299b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.147707 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.147739 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a909008-f754-49fb-87cb-dc0d09c91bfd-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.147753 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.152612 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-scripts" (OuterVolumeSpecName: "scripts") pod "e8331eb7-653b-4a6c-ada8-f13dcf916a8a" (UID: "e8331eb7-653b-4a6c-ada8-f13dcf916a8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.152892 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-config-data" (OuterVolumeSpecName: "config-data") pod "e8331eb7-653b-4a6c-ada8-f13dcf916a8a" (UID: "e8331eb7-653b-4a6c-ada8-f13dcf916a8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.187897 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4e56f804-b6ae-4d86-8e0b-9aed08a299b4" (UID: "4e56f804-b6ae-4d86-8e0b-9aed08a299b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.250663 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.250706 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.250720 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8331eb7-653b-4a6c-ada8-f13dcf916a8a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.252570 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-config" (OuterVolumeSpecName: "config") pod "4e56f804-b6ae-4d86-8e0b-9aed08a299b4" (UID: "4e56f804-b6ae-4d86-8e0b-9aed08a299b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.291892 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4e56f804-b6ae-4d86-8e0b-9aed08a299b4" (UID: "4e56f804-b6ae-4d86-8e0b-9aed08a299b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.293638 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4e56f804-b6ae-4d86-8e0b-9aed08a299b4" (UID: "4e56f804-b6ae-4d86-8e0b-9aed08a299b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.353349 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.353384 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.353394 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e56f804-b6ae-4d86-8e0b-9aed08a299b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.540355 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-bcn2w"] Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.541117 4715 scope.go:117] "RemoveContainer" containerID="bdd78e6d8ce2e391a2a3c679b069dc2b9e4dc37f274bf1dbe8a52828ada6e4b2" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.701863 4715 scope.go:117] "RemoveContainer" containerID="8f255ea6ba69fe26492e6e42a0adcf97b1c6441c697666999c812155e5eb0e6d" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.753463 4715 scope.go:117] "RemoveContainer" containerID="207bb21cc12f37dde2b5a2635bd91b784400406326add9b1bbfa01c6fe3a1eb4" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.856170 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.888302 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6dc6d96-gj9d4" event={"ID":"aad8392c-fc50-4cd5-af7c-9d186021d336","Type":"ContainerStarted","Data":"b7ac3f72224434537c9126cb8c6d36a9ad06e081b20a668f4c4732bf75ab85e6"} Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.913391 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57458fcb8f-w827q" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.921827 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5978d7d969-7vwfh" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.922722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" event={"ID":"decf7e1e-62c0-4bb0-8e6e-179d07821e4f","Type":"ContainerStarted","Data":"c88cf90fc86be8c27cf863e402670dda4137ed44ea5bc414e1df9442c4371f61"} Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.922792 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8599dd67c5-qs56p" Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.923359 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-central-agent" containerID="cri-o://de12a58b72d05fca1fe302798225bc3ba3828024703f7b8f60ac4c0a898741c7" gracePeriod=30 Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.923489 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="sg-core" containerID="cri-o://fda30309ef0e04fbb9e89c36dd6a42cefd5ebbde0e791d60f1a6d2d7b742a8d3" gracePeriod=30 Dec 04 14:19:35 crc kubenswrapper[4715]: I1204 14:19:35.923541 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-notification-agent" containerID="cri-o://784803b676bc40f74c8cb814c6ded8ae84d6a7c073a2a93c40de59bd325ab990" gracePeriod=30 Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.041860 4715 scope.go:117] "RemoveContainer" containerID="dbb5250430bbee33218c769d14a9d9d05e1495c461ecb329525091ccaaed8b15" Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.058011 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.098936 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z4zqf"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.133435 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z4zqf"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.204855 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5978d7d969-7vwfh"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.268110 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5978d7d969-7vwfh"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.281320 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8599dd67c5-qs56p"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.293204 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8599dd67c5-qs56p"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.303485 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57458fcb8f-w827q"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.312091 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-57458fcb8f-w827q"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.328754 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d7bd77bfb-nlk9m"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.341181 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-swb86"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.568472 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.780376 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.954753 4715 generic.go:334] "Generic (PLEG): container finished" podID="2773b750-4db6-46e8-9337-4fd592cc8410" containerID="31b336103582571f529a66fd90896aefebf375cbf786c57d0089b2fe75d25b45" exitCode=0 Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.954867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7896b66d76-jtzqq" event={"ID":"2773b750-4db6-46e8-9337-4fd592cc8410","Type":"ContainerDied","Data":"31b336103582571f529a66fd90896aefebf375cbf786c57d0089b2fe75d25b45"} Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.956727 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c630f1e1-b605-4282-b136-647528bc86b4","Type":"ContainerStarted","Data":"15cfabd30b29d5fa2b69608598802bd8b527d711eee037c5cf3ce4ce7e3bc8d1"} Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.960324 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79b4bf01-31b0-4a13-9359-8574e4fba640","Type":"ContainerStarted","Data":"7e0c104cc881c4a7a3018ff021c6f29452c5b8cec651df7b928e1a8dc50f32d1"} Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.973048 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7bd77bfb-nlk9m" event={"ID":"d49eba49-c182-4ec6-9a61-d14cddf541fd","Type":"ContainerStarted","Data":"32d6029c149005d6daaf527ea08fe637a30db48ac7d8123e46d08170e7274461"} Dec 04 14:19:36 crc kubenswrapper[4715]: I1204 14:19:36.978376 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" event={"ID":"38ade7f7-11d1-43a2-9520-f6b217314df9","Type":"ContainerStarted","Data":"06274e7af106a34cfe4876fb1cfe115e4432fbef3916f885debf18daa295b32d"} Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:36.996883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" event={"ID":"decf7e1e-62c0-4bb0-8e6e-179d07821e4f","Type":"ContainerStarted","Data":"33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c"} Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:36.997123 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" podUID="decf7e1e-62c0-4bb0-8e6e-179d07821e4f" containerName="init" containerID="cri-o://33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c" gracePeriod=10 Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.000487 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerID="fda30309ef0e04fbb9e89c36dd6a42cefd5ebbde0e791d60f1a6d2d7b742a8d3" exitCode=2 Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.000513 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerID="de12a58b72d05fca1fe302798225bc3ba3828024703f7b8f60ac4c0a898741c7" exitCode=0 Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.000534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerDied","Data":"fda30309ef0e04fbb9e89c36dd6a42cefd5ebbde0e791d60f1a6d2d7b742a8d3"} Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.000566 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerDied","Data":"de12a58b72d05fca1fe302798225bc3ba3828024703f7b8f60ac4c0a898741c7"} Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.211941 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" path="/var/lib/kubelet/pods/0a909008-f754-49fb-87cb-dc0d09c91bfd/volumes" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.212775 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" path="/var/lib/kubelet/pods/4e56f804-b6ae-4d86-8e0b-9aed08a299b4/volumes" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.214453 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" path="/var/lib/kubelet/pods/a1c0859d-f9ae-457f-9c44-ea9891d7756b/volumes" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.215859 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" path="/var/lib/kubelet/pods/e8331eb7-653b-4a6c-ada8-f13dcf916a8a/volumes" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.646515 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.760652 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2773b750-4db6-46e8-9337-4fd592cc8410-logs\") pod \"2773b750-4db6-46e8-9337-4fd592cc8410\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.761175 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9s95\" (UniqueName: \"kubernetes.io/projected/2773b750-4db6-46e8-9337-4fd592cc8410-kube-api-access-x9s95\") pod \"2773b750-4db6-46e8-9337-4fd592cc8410\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.766951 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2773b750-4db6-46e8-9337-4fd592cc8410-logs" (OuterVolumeSpecName: "logs") pod "2773b750-4db6-46e8-9337-4fd592cc8410" (UID: "2773b750-4db6-46e8-9337-4fd592cc8410"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.768816 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data\") pod \"2773b750-4db6-46e8-9337-4fd592cc8410\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.768850 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data-custom\") pod \"2773b750-4db6-46e8-9337-4fd592cc8410\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.768890 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-combined-ca-bundle\") pod \"2773b750-4db6-46e8-9337-4fd592cc8410\" (UID: \"2773b750-4db6-46e8-9337-4fd592cc8410\") " Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.769352 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2773b750-4db6-46e8-9337-4fd592cc8410-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.779591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2773b750-4db6-46e8-9337-4fd592cc8410" (UID: "2773b750-4db6-46e8-9337-4fd592cc8410"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.789154 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2773b750-4db6-46e8-9337-4fd592cc8410-kube-api-access-x9s95" (OuterVolumeSpecName: "kube-api-access-x9s95") pod "2773b750-4db6-46e8-9337-4fd592cc8410" (UID: "2773b750-4db6-46e8-9337-4fd592cc8410"). InnerVolumeSpecName "kube-api-access-x9s95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.876232 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.876264 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9s95\" (UniqueName: \"kubernetes.io/projected/2773b750-4db6-46e8-9337-4fd592cc8410-kube-api-access-x9s95\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.881376 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2773b750-4db6-46e8-9337-4fd592cc8410" (UID: "2773b750-4db6-46e8-9337-4fd592cc8410"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.977878 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:37 crc kubenswrapper[4715]: I1204 14:19:37.993706 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data" (OuterVolumeSpecName: "config-data") pod "2773b750-4db6-46e8-9337-4fd592cc8410" (UID: "2773b750-4db6-46e8-9337-4fd592cc8410"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.030179 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6dc6d96-gj9d4" event={"ID":"aad8392c-fc50-4cd5-af7c-9d186021d336","Type":"ContainerStarted","Data":"3fe498a00685ae30e697ec51ab5aeed4e9d98ae50be83e46d46368b9429ebedd"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.033738 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.033912 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.038449 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.038945 4715 generic.go:334] "Generic (PLEG): container finished" podID="decf7e1e-62c0-4bb0-8e6e-179d07821e4f" containerID="33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c" exitCode=0 Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.039006 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" event={"ID":"decf7e1e-62c0-4bb0-8e6e-179d07821e4f","Type":"ContainerDied","Data":"33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.039049 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" event={"ID":"decf7e1e-62c0-4bb0-8e6e-179d07821e4f","Type":"ContainerDied","Data":"c88cf90fc86be8c27cf863e402670dda4137ed44ea5bc414e1df9442c4371f61"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.039076 4715 scope.go:117] "RemoveContainer" containerID="33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.046365 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7896b66d76-jtzqq" event={"ID":"2773b750-4db6-46e8-9337-4fd592cc8410","Type":"ContainerDied","Data":"ef9376e0b8cade422a16d95bfb469628f9a2f44e7099cd517d398627003e7e58"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.046483 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7896b66d76-jtzqq" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.074524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7bd77bfb-nlk9m" event={"ID":"d49eba49-c182-4ec6-9a61-d14cddf541fd","Type":"ContainerStarted","Data":"0f42ca9a061f45d393889a4cfa54ca235d7648375ad5d892f8dc431de02d3d95"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.074583 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7bd77bfb-nlk9m" event={"ID":"d49eba49-c182-4ec6-9a61-d14cddf541fd","Type":"ContainerStarted","Data":"dd7b9e026fbce19f88bd2063857eba783cb3f3f66bde675059e0b4d466564778"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.075631 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.080544 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h2cb\" (UniqueName: \"kubernetes.io/projected/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-kube-api-access-5h2cb\") pod \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.080627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-nb\") pod \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.080680 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-config\") pod \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.080728 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-svc\") pod \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.080760 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-swift-storage-0\") pod \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.080916 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-sb\") pod \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\" (UID: \"decf7e1e-62c0-4bb0-8e6e-179d07821e4f\") " Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.081700 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2773b750-4db6-46e8-9337-4fd592cc8410-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.087406 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-kube-api-access-5h2cb" (OuterVolumeSpecName: "kube-api-access-5h2cb") pod "decf7e1e-62c0-4bb0-8e6e-179d07821e4f" (UID: "decf7e1e-62c0-4bb0-8e6e-179d07821e4f"). InnerVolumeSpecName "kube-api-access-5h2cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.099073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" event={"ID":"38ade7f7-11d1-43a2-9520-f6b217314df9","Type":"ContainerStarted","Data":"2352bbb6dcce27827a0cd26ec0464d0f9b19c3beb79e8f8861956afcf3713f86"} Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.117686 4715 scope.go:117] "RemoveContainer" containerID="33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c" Dec 04 14:19:38 crc kubenswrapper[4715]: E1204 14:19:38.120197 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c\": container with ID starting with 33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c not found: ID does not exist" containerID="33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.120243 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c"} err="failed to get container status \"33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c\": rpc error: code = NotFound desc = could not find container \"33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c\": container with ID starting with 33953d2f5434c7fe28d9eb2d7c5301db2bf54d059897144d8a7760737083432c not found: ID does not exist" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.120271 4715 scope.go:117] "RemoveContainer" containerID="31b336103582571f529a66fd90896aefebf375cbf786c57d0089b2fe75d25b45" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.138828 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65f6dc6d96-gj9d4" podStartSLOduration=14.13878303 podStartE2EDuration="14.13878303s" podCreationTimestamp="2025-12-04 14:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:38.075954619 +0000 UTC m=+1335.144672834" watchObservedRunningTime="2025-12-04 14:19:38.13878303 +0000 UTC m=+1335.207501245" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.153849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "decf7e1e-62c0-4bb0-8e6e-179d07821e4f" (UID: "decf7e1e-62c0-4bb0-8e6e-179d07821e4f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.158781 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "decf7e1e-62c0-4bb0-8e6e-179d07821e4f" (UID: "decf7e1e-62c0-4bb0-8e6e-179d07821e4f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.163165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "decf7e1e-62c0-4bb0-8e6e-179d07821e4f" (UID: "decf7e1e-62c0-4bb0-8e6e-179d07821e4f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.166237 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d7bd77bfb-nlk9m" podStartSLOduration=5.166214513 podStartE2EDuration="5.166214513s" podCreationTimestamp="2025-12-04 14:19:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:38.11917907 +0000 UTC m=+1335.187897305" watchObservedRunningTime="2025-12-04 14:19:38.166214513 +0000 UTC m=+1335.234932718" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.166725 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "decf7e1e-62c0-4bb0-8e6e-179d07821e4f" (UID: "decf7e1e-62c0-4bb0-8e6e-179d07821e4f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.167990 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-config" (OuterVolumeSpecName: "config") pod "decf7e1e-62c0-4bb0-8e6e-179d07821e4f" (UID: "decf7e1e-62c0-4bb0-8e6e-179d07821e4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.185973 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h2cb\" (UniqueName: \"kubernetes.io/projected/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-kube-api-access-5h2cb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.186016 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.186051 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.186064 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.186075 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.186087 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/decf7e1e-62c0-4bb0-8e6e-179d07821e4f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.194104 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7896b66d76-jtzqq"] Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.219466 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7896b66d76-jtzqq"] Dec 04 14:19:38 crc kubenswrapper[4715]: I1204 14:19:38.387272 4715 scope.go:117] "RemoveContainer" containerID="18db19ae897873d8198e1589cd364976097faa3820993571a28b5385fc8b11d6" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.155394 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-bcn2w" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.172590 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerID="784803b676bc40f74c8cb814c6ded8ae84d6a7c073a2a93c40de59bd325ab990" exitCode=0 Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.172673 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerDied","Data":"784803b676bc40f74c8cb814c6ded8ae84d6a7c073a2a93c40de59bd325ab990"} Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.222236 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" path="/var/lib/kubelet/pods/2773b750-4db6-46e8-9337-4fd592cc8410/volumes" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.224806 4715 generic.go:334] "Generic (PLEG): container finished" podID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerID="2352bbb6dcce27827a0cd26ec0464d0f9b19c3beb79e8f8861956afcf3713f86" exitCode=0 Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.225793 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c630f1e1-b605-4282-b136-647528bc86b4","Type":"ContainerStarted","Data":"c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5"} Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.225820 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79b4bf01-31b0-4a13-9359-8574e4fba640","Type":"ContainerStarted","Data":"b44351b04a9829963f43cdd221275304338a9a87f7d6d2d1b9dbadd920e0567b"} Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.225833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" event={"ID":"38ade7f7-11d1-43a2-9520-f6b217314df9","Type":"ContainerDied","Data":"2352bbb6dcce27827a0cd26ec0464d0f9b19c3beb79e8f8861956afcf3713f86"} Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.225848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" event={"ID":"38ade7f7-11d1-43a2-9520-f6b217314df9","Type":"ContainerStarted","Data":"c7a72ccf3f913c5adcbaad23be5746a33b02438ef6228a4f3e7dfaa38f1d10d0"} Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.226329 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.259201 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-bcn2w"] Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.286889 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-bcn2w"] Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.315304 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" podStartSLOduration=5.315281333 podStartE2EDuration="5.315281333s" podCreationTimestamp="2025-12-04 14:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:39.304071689 +0000 UTC m=+1336.372789904" watchObservedRunningTime="2025-12-04 14:19:39.315281333 +0000 UTC m=+1336.383999548" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.562162 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655390 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-log-httpd\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655445 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-run-httpd\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655471 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-combined-ca-bundle\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655502 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-config-data\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655537 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqtrm\" (UniqueName: \"kubernetes.io/projected/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-kube-api-access-tqtrm\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655572 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-scripts\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.655681 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-sg-core-conf-yaml\") pod \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\" (UID: \"8b94f46e-2d75-479c-9ef6-b8fac3332e2f\") " Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.656565 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.657748 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.671440 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-kube-api-access-tqtrm" (OuterVolumeSpecName: "kube-api-access-tqtrm") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "kube-api-access-tqtrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.677443 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-scripts" (OuterVolumeSpecName: "scripts") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.706245 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.731170 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-z4zqf" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.746152 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-config-data" (OuterVolumeSpecName: "config-data") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.759171 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.759214 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.759226 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.759238 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqtrm\" (UniqueName: \"kubernetes.io/projected/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-kube-api-access-tqtrm\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.759250 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.759258 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.774203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b94f46e-2d75-479c-9ef6-b8fac3332e2f" (UID: "8b94f46e-2d75-479c-9ef6-b8fac3332e2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:39 crc kubenswrapper[4715]: I1204 14:19:39.860812 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b94f46e-2d75-479c-9ef6-b8fac3332e2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.247159 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-58655b5d9c-m5lkw"] Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.247960 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decf7e1e-62c0-4bb0-8e6e-179d07821e4f" containerName="init" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.247978 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="decf7e1e-62c0-4bb0-8e6e-179d07821e4f" containerName="init" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248000 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="init" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248009 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="init" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248048 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="sg-core" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248056 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="sg-core" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248070 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248078 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248093 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248101 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248117 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248124 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248144 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-central-agent" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248152 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-central-agent" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248167 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248174 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248184 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248192 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api-log" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248207 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248214 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248228 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-notification-agent" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248236 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-notification-agent" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248246 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248254 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248267 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="dnsmasq-dns" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248275 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="dnsmasq-dns" Dec 04 14:19:40 crc kubenswrapper[4715]: E1204 14:19:40.248293 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248302 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248584 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-notification-agent" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248597 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248607 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="ceilometer-central-agent" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248623 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248637 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248650 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a909008-f754-49fb-87cb-dc0d09c91bfd" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248662 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248675 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248687 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" containerName="sg-core" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248698 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e56f804-b6ae-4d86-8e0b-9aed08a299b4" containerName="dnsmasq-dns" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248707 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="decf7e1e-62c0-4bb0-8e6e-179d07821e4f" containerName="init" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248719 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c0859d-f9ae-457f-9c44-ea9891d7756b" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.248729 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8331eb7-653b-4a6c-ada8-f13dcf916a8a" containerName="horizon-log" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.249974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.259270 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.259389 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.274392 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58655b5d9c-m5lkw"] Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.289081 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8b94f46e-2d75-479c-9ef6-b8fac3332e2f","Type":"ContainerDied","Data":"2e7948acdab947fccfea2245893ee375640cadd491d806f2745ab1dd2937798c"} Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.289138 4715 scope.go:117] "RemoveContainer" containerID="fda30309ef0e04fbb9e89c36dd6a42cefd5ebbde0e791d60f1a6d2d7b742a8d3" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.289299 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.309228 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c630f1e1-b605-4282-b136-647528bc86b4","Type":"ContainerStarted","Data":"be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296"} Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.318832 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api-log" containerID="cri-o://b44351b04a9829963f43cdd221275304338a9a87f7d6d2d1b9dbadd920e0567b" gracePeriod=30 Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.319213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79b4bf01-31b0-4a13-9359-8574e4fba640","Type":"ContainerStarted","Data":"9591a75565166e919304c0389e032ae1a259b9bf26010ef1316121b6822b7bd5"} Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.320186 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api" containerID="cri-o://9591a75565166e919304c0389e032ae1a259b9bf26010ef1316121b6822b7bd5" gracePeriod=30 Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.320374 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.343209 4715 scope.go:117] "RemoveContainer" containerID="784803b676bc40f74c8cb814c6ded8ae84d6a7c073a2a93c40de59bd325ab990" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.371980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-ovndb-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.372182 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-config\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.372248 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-public-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.372317 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-httpd-config\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.372407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc95x\" (UniqueName: \"kubernetes.io/projected/4e9c813e-1ad1-40fd-b015-a312efff6d9e-kube-api-access-cc95x\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.372431 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-internal-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.372485 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-combined-ca-bundle\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.392094 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.406401 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.410015 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.409990828 podStartE2EDuration="6.409990828s" podCreationTimestamp="2025-12-04 14:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:40.384983741 +0000 UTC m=+1337.453701976" watchObservedRunningTime="2025-12-04 14:19:40.409990828 +0000 UTC m=+1337.478709043" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.433198 4715 scope.go:117] "RemoveContainer" containerID="de12a58b72d05fca1fe302798225bc3ba3828024703f7b8f60ac4c0a898741c7" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474481 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-ovndb-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474620 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-config\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474677 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-public-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474731 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-httpd-config\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474777 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc95x\" (UniqueName: \"kubernetes.io/projected/4e9c813e-1ad1-40fd-b015-a312efff6d9e-kube-api-access-cc95x\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-internal-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.474830 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-combined-ca-bundle\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.481597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-internal-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.483366 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-config\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.484054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-httpd-config\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.486878 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-combined-ca-bundle\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.489817 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-ovndb-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.493463 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.495796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9c813e-1ad1-40fd-b015-a312efff6d9e-public-tls-certs\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.512336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc95x\" (UniqueName: \"kubernetes.io/projected/4e9c813e-1ad1-40fd-b015-a312efff6d9e-kube-api-access-cc95x\") pod \"neutron-58655b5d9c-m5lkw\" (UID: \"4e9c813e-1ad1-40fd-b015-a312efff6d9e\") " pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.516589 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.283811269 podStartE2EDuration="7.516551864s" podCreationTimestamp="2025-12-04 14:19:33 +0000 UTC" firstStartedPulling="2025-12-04 14:19:36.159317703 +0000 UTC m=+1333.228035918" lastFinishedPulling="2025-12-04 14:19:37.392058298 +0000 UTC m=+1334.460776513" observedRunningTime="2025-12-04 14:19:40.419775313 +0000 UTC m=+1337.488493548" watchObservedRunningTime="2025-12-04 14:19:40.516551864 +0000 UTC m=+1337.585270089" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.520824 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.525388 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.526889 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577597 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-run-httpd\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-config-data\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577762 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wwbt\" (UniqueName: \"kubernetes.io/projected/268793f0-3a54-4a16-a059-7573491f9ef5-kube-api-access-9wwbt\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577817 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-scripts\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577840 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577881 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.577936 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-log-httpd\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.578073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.583441 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.679224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-run-httpd\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.679588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-config-data\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.679668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wwbt\" (UniqueName: \"kubernetes.io/projected/268793f0-3a54-4a16-a059-7573491f9ef5-kube-api-access-9wwbt\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.679732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-scripts\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.679756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.679950 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.680024 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-log-httpd\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.680114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-run-httpd\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.683427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-log-httpd\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.689279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.689543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-config-data\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.696645 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-scripts\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.721475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.725543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wwbt\" (UniqueName: \"kubernetes.io/projected/268793f0-3a54-4a16-a059-7573491f9ef5-kube-api-access-9wwbt\") pod \"ceilometer-0\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " pod="openstack/ceilometer-0" Dec 04 14:19:40 crc kubenswrapper[4715]: I1204 14:19:40.913731 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.197759 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b94f46e-2d75-479c-9ef6-b8fac3332e2f" path="/var/lib/kubelet/pods/8b94f46e-2d75-479c-9ef6-b8fac3332e2f/volumes" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.198808 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="decf7e1e-62c0-4bb0-8e6e-179d07821e4f" path="/var/lib/kubelet/pods/decf7e1e-62c0-4bb0-8e6e-179d07821e4f/volumes" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.346840 4715 generic.go:334] "Generic (PLEG): container finished" podID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerID="9591a75565166e919304c0389e032ae1a259b9bf26010ef1316121b6822b7bd5" exitCode=0 Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.346883 4715 generic.go:334] "Generic (PLEG): container finished" podID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerID="b44351b04a9829963f43cdd221275304338a9a87f7d6d2d1b9dbadd920e0567b" exitCode=143 Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.346944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79b4bf01-31b0-4a13-9359-8574e4fba640","Type":"ContainerDied","Data":"9591a75565166e919304c0389e032ae1a259b9bf26010ef1316121b6822b7bd5"} Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.346978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79b4bf01-31b0-4a13-9359-8574e4fba640","Type":"ContainerDied","Data":"b44351b04a9829963f43cdd221275304338a9a87f7d6d2d1b9dbadd920e0567b"} Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.346991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79b4bf01-31b0-4a13-9359-8574e4fba640","Type":"ContainerDied","Data":"7e0c104cc881c4a7a3018ff021c6f29452c5b8cec651df7b928e1a8dc50f32d1"} Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.347003 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e0c104cc881c4a7a3018ff021c6f29452c5b8cec651df7b928e1a8dc50f32d1" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.383588 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58655b5d9c-m5lkw"] Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.390375 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502269 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-combined-ca-bundle\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502342 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data-custom\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502386 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502452 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79b4bf01-31b0-4a13-9359-8574e4fba640-etc-machine-id\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502476 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-scripts\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502505 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b4bf01-31b0-4a13-9359-8574e4fba640-logs\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.502629 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxd9s\" (UniqueName: \"kubernetes.io/projected/79b4bf01-31b0-4a13-9359-8574e4fba640-kube-api-access-vxd9s\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.508263 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79b4bf01-31b0-4a13-9359-8574e4fba640-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.513610 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79b4bf01-31b0-4a13-9359-8574e4fba640-logs" (OuterVolumeSpecName: "logs") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.523172 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-scripts" (OuterVolumeSpecName: "scripts") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.523443 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b4bf01-31b0-4a13-9359-8574e4fba640-kube-api-access-vxd9s" (OuterVolumeSpecName: "kube-api-access-vxd9s") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "kube-api-access-vxd9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.532165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.574204 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.577800 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.605763 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data" (OuterVolumeSpecName: "config-data") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.605968 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data\") pod \"79b4bf01-31b0-4a13-9359-8574e4fba640\" (UID: \"79b4bf01-31b0-4a13-9359-8574e4fba640\") " Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606570 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxd9s\" (UniqueName: \"kubernetes.io/projected/79b4bf01-31b0-4a13-9359-8574e4fba640-kube-api-access-vxd9s\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606589 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606602 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: W1204 14:19:41.606596 4715 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/79b4bf01-31b0-4a13-9359-8574e4fba640/volumes/kubernetes.io~secret/config-data Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606614 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79b4bf01-31b0-4a13-9359-8574e4fba640-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606625 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606624 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data" (OuterVolumeSpecName: "config-data") pod "79b4bf01-31b0-4a13-9359-8574e4fba640" (UID: "79b4bf01-31b0-4a13-9359-8574e4fba640"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.606640 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b4bf01-31b0-4a13-9359-8574e4fba640-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.708409 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b4bf01-31b0-4a13-9359-8574e4fba640-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:41 crc kubenswrapper[4715]: I1204 14:19:41.934716 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.028565 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7896b66d76-jtzqq" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.028666 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7896b66d76-jtzqq" podUID="2773b750-4db6-46e8-9337-4fd592cc8410" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.368752 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerStarted","Data":"69e8a174021e90fc8244cb678755a51343fa78346c8ca9e1b8fc2cee10998bc9"} Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.368811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerStarted","Data":"9a27e84cb32ef373108f2f05b949121de6303d1f04893ed036248d3604ec1676"} Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.378698 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.379223 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58655b5d9c-m5lkw" event={"ID":"4e9c813e-1ad1-40fd-b015-a312efff6d9e","Type":"ContainerStarted","Data":"60bcf66ffe652674e95c391165b004d72411d83cb9bfbe4eb4a26e03f391b8f4"} Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.379288 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58655b5d9c-m5lkw" event={"ID":"4e9c813e-1ad1-40fd-b015-a312efff6d9e","Type":"ContainerStarted","Data":"b6fa30c59cfa8673d8b4027e4ea3ea2081508d8f1efd7ca4c1834fc76cdb38fc"} Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.379304 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58655b5d9c-m5lkw" event={"ID":"4e9c813e-1ad1-40fd-b015-a312efff6d9e","Type":"ContainerStarted","Data":"337950a8556b7fa4de412d46751cc719592e1736a36fedba3bfd1f193c2ad0e3"} Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.380292 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.527539 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-58655b5d9c-m5lkw" podStartSLOduration=2.527519455 podStartE2EDuration="2.527519455s" podCreationTimestamp="2025-12-04 14:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:42.520192577 +0000 UTC m=+1339.588910812" watchObservedRunningTime="2025-12-04 14:19:42.527519455 +0000 UTC m=+1339.596237670" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.561133 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.580918 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.596615 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:42 crc kubenswrapper[4715]: E1204 14:19:42.597691 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.597710 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api" Dec 04 14:19:42 crc kubenswrapper[4715]: E1204 14:19:42.597722 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api-log" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.597738 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api-log" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.598271 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api-log" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.598332 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" containerName="cinder-api" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.603380 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.612260 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.612575 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.615686 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.656914 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.657345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.657590 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87853446-d88c-4d46-9fea-ff90af5617ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.657697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgt8b\" (UniqueName: \"kubernetes.io/projected/87853446-d88c-4d46-9fea-ff90af5617ba-kube-api-access-qgt8b\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.657813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-scripts\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.657913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87853446-d88c-4d46-9fea-ff90af5617ba-logs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.657997 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.658334 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.658454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-config-data\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.664471 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761139 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-scripts\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761528 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87853446-d88c-4d46-9fea-ff90af5617ba-logs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761570 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761667 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-config-data\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761758 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87853446-d88c-4d46-9fea-ff90af5617ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.761873 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgt8b\" (UniqueName: \"kubernetes.io/projected/87853446-d88c-4d46-9fea-ff90af5617ba-kube-api-access-qgt8b\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.768573 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.773711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-scripts\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.774008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87853446-d88c-4d46-9fea-ff90af5617ba-logs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.776081 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-config-data\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.776171 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87853446-d88c-4d46-9fea-ff90af5617ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.777258 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.779155 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.788909 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87853446-d88c-4d46-9fea-ff90af5617ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.795903 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgt8b\" (UniqueName: \"kubernetes.io/projected/87853446-d88c-4d46-9fea-ff90af5617ba-kube-api-access-qgt8b\") pod \"cinder-api-0\" (UID: \"87853446-d88c-4d46-9fea-ff90af5617ba\") " pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.940545 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:19:42 crc kubenswrapper[4715]: I1204 14:19:42.943698 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:19:43 crc kubenswrapper[4715]: I1204 14:19:43.544654 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79b4bf01-31b0-4a13-9359-8574e4fba640" path="/var/lib/kubelet/pods/79b4bf01-31b0-4a13-9359-8574e4fba640/volumes" Dec 04 14:19:43 crc kubenswrapper[4715]: I1204 14:19:43.557093 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.028597 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.325201 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.346437 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65f6dc6d96-gj9d4" Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.440328 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerStarted","Data":"0d909ded27468e80eef0dad7f263f1a9c7f08e4812a144ca6407d3a5d0cd228c"} Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.474534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87853446-d88c-4d46-9fea-ff90af5617ba","Type":"ContainerStarted","Data":"25aea46f5fb4b66548d7af6d646b348f1831bbc710609b8ac2ccd943b1575629"} Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.845659 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.850424 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.948233 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-j6rz7"] Dec 04 14:19:44 crc kubenswrapper[4715]: I1204 14:19:44.948663 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" containerName="dnsmasq-dns" containerID="cri-o://4249d77a08662aebb6db84ab2ea51b79e5a7ea600284207d7c0c9160afb94728" gracePeriod=10 Dec 04 14:19:45 crc kubenswrapper[4715]: I1204 14:19:45.038847 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:45 crc kubenswrapper[4715]: I1204 14:19:45.485990 4715 generic.go:334] "Generic (PLEG): container finished" podID="92a3bae3-f661-492c-bb10-cf636c135d06" containerID="4249d77a08662aebb6db84ab2ea51b79e5a7ea600284207d7c0c9160afb94728" exitCode=0 Dec 04 14:19:45 crc kubenswrapper[4715]: I1204 14:19:45.486076 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" event={"ID":"92a3bae3-f661-492c-bb10-cf636c135d06","Type":"ContainerDied","Data":"4249d77a08662aebb6db84ab2ea51b79e5a7ea600284207d7c0c9160afb94728"} Dec 04 14:19:45 crc kubenswrapper[4715]: I1204 14:19:45.488613 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="cinder-scheduler" containerID="cri-o://c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5" gracePeriod=30 Dec 04 14:19:45 crc kubenswrapper[4715]: I1204 14:19:45.488984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerStarted","Data":"321883fe0907f52c898143a19c0d0b867315c6dc23c0e187f48da043d04db650"} Dec 04 14:19:45 crc kubenswrapper[4715]: I1204 14:19:45.489129 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="probe" containerID="cri-o://be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296" gracePeriod=30 Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.519282 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" event={"ID":"92a3bae3-f661-492c-bb10-cf636c135d06","Type":"ContainerDied","Data":"067e8a0348f2a1d1fdf741f1bf0c2edf5a0b3ce12c0913a72a413bbc3db9554f"} Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.519799 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="067e8a0348f2a1d1fdf741f1bf0c2edf5a0b3ce12c0913a72a413bbc3db9554f" Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.530528 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87853446-d88c-4d46-9fea-ff90af5617ba","Type":"ContainerStarted","Data":"af5ad280f19b6065f0290edb04c84e8c867496ac5c62c267e5f1e6e20a8192dc"} Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.565253 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.658929 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5b79f64b59-wfklg" Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.829984 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-sb\") pod \"92a3bae3-f661-492c-bb10-cf636c135d06\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.833509 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k52s\" (UniqueName: \"kubernetes.io/projected/92a3bae3-f661-492c-bb10-cf636c135d06-kube-api-access-7k52s\") pod \"92a3bae3-f661-492c-bb10-cf636c135d06\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.834454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-config\") pod \"92a3bae3-f661-492c-bb10-cf636c135d06\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.834614 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-nb\") pod \"92a3bae3-f661-492c-bb10-cf636c135d06\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.834652 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-swift-storage-0\") pod \"92a3bae3-f661-492c-bb10-cf636c135d06\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.834672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-svc\") pod \"92a3bae3-f661-492c-bb10-cf636c135d06\" (UID: \"92a3bae3-f661-492c-bb10-cf636c135d06\") " Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.847441 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a3bae3-f661-492c-bb10-cf636c135d06-kube-api-access-7k52s" (OuterVolumeSpecName: "kube-api-access-7k52s") pod "92a3bae3-f661-492c-bb10-cf636c135d06" (UID: "92a3bae3-f661-492c-bb10-cf636c135d06"). InnerVolumeSpecName "kube-api-access-7k52s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.859480 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5b798c69c6-g6fvm" Dec 04 14:19:46 crc kubenswrapper[4715]: I1204 14:19:46.936436 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k52s\" (UniqueName: \"kubernetes.io/projected/92a3bae3-f661-492c-bb10-cf636c135d06-kube-api-access-7k52s\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.006927 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-754dbdc69b-nwpzv"] Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.007470 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon-log" containerID="cri-o://1341f08949be200925226253e4ab03bfe80d9e1c59dcf0907a757e3dab2e4ec3" gracePeriod=30 Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.007599 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" containerID="cri-o://f1297a46867ceb825461ae2c754dc4d5b3ba962c40fa5376fd6f94ccf7879eb0" gracePeriod=30 Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.051428 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.132867 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92a3bae3-f661-492c-bb10-cf636c135d06" (UID: "92a3bae3-f661-492c-bb10-cf636c135d06"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.158409 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.371312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "92a3bae3-f661-492c-bb10-cf636c135d06" (UID: "92a3bae3-f661-492c-bb10-cf636c135d06"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.380745 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "92a3bae3-f661-492c-bb10-cf636c135d06" (UID: "92a3bae3-f661-492c-bb10-cf636c135d06"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.397218 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92a3bae3-f661-492c-bb10-cf636c135d06" (UID: "92a3bae3-f661-492c-bb10-cf636c135d06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.407058 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-config" (OuterVolumeSpecName: "config") pod "92a3bae3-f661-492c-bb10-cf636c135d06" (UID: "92a3bae3-f661-492c-bb10-cf636c135d06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.547774 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.547811 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.547823 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.547834 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a3bae3-f661-492c-bb10-cf636c135d06-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.565640 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerStarted","Data":"565247e23769c9075e305bf49b769a230bbb448dc04958fdb6dd922da3393a8c"} Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.567197 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.574673 4715 generic.go:334] "Generic (PLEG): container finished" podID="c630f1e1-b605-4282-b136-647528bc86b4" containerID="be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296" exitCode=0 Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.574775 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-j6rz7" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.578568 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c630f1e1-b605-4282-b136-647528bc86b4","Type":"ContainerDied","Data":"be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296"} Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.602536 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.776690812 podStartE2EDuration="7.602513623s" podCreationTimestamp="2025-12-04 14:19:40 +0000 UTC" firstStartedPulling="2025-12-04 14:19:41.608291441 +0000 UTC m=+1338.677009666" lastFinishedPulling="2025-12-04 14:19:46.434114252 +0000 UTC m=+1343.502832477" observedRunningTime="2025-12-04 14:19:47.587583029 +0000 UTC m=+1344.656301294" watchObservedRunningTime="2025-12-04 14:19:47.602513623 +0000 UTC m=+1344.671231838" Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.628101 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-j6rz7"] Dec 04 14:19:47 crc kubenswrapper[4715]: I1204 14:19:47.635283 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-j6rz7"] Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.461146 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.490703 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-scripts\") pod \"c630f1e1-b605-4282-b136-647528bc86b4\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.490799 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-combined-ca-bundle\") pod \"c630f1e1-b605-4282-b136-647528bc86b4\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.496839 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-scripts" (OuterVolumeSpecName: "scripts") pod "c630f1e1-b605-4282-b136-647528bc86b4" (UID: "c630f1e1-b605-4282-b136-647528bc86b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.584787 4715 generic.go:334] "Generic (PLEG): container finished" podID="c630f1e1-b605-4282-b136-647528bc86b4" containerID="c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5" exitCode=0 Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.584844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c630f1e1-b605-4282-b136-647528bc86b4","Type":"ContainerDied","Data":"c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5"} Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.584869 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c630f1e1-b605-4282-b136-647528bc86b4","Type":"ContainerDied","Data":"15cfabd30b29d5fa2b69608598802bd8b527d711eee037c5cf3ce4ce7e3bc8d1"} Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.584887 4715 scope.go:117] "RemoveContainer" containerID="be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.585065 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.592474 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c630f1e1-b605-4282-b136-647528bc86b4" (UID: "c630f1e1-b605-4282-b136-647528bc86b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.593266 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dznxb\" (UniqueName: \"kubernetes.io/projected/c630f1e1-b605-4282-b136-647528bc86b4-kube-api-access-dznxb\") pod \"c630f1e1-b605-4282-b136-647528bc86b4\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.593340 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data-custom\") pod \"c630f1e1-b605-4282-b136-647528bc86b4\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.593401 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c630f1e1-b605-4282-b136-647528bc86b4-etc-machine-id\") pod \"c630f1e1-b605-4282-b136-647528bc86b4\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.593428 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data\") pod \"c630f1e1-b605-4282-b136-647528bc86b4\" (UID: \"c630f1e1-b605-4282-b136-647528bc86b4\") " Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.593874 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.593894 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.594431 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c630f1e1-b605-4282-b136-647528bc86b4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c630f1e1-b605-4282-b136-647528bc86b4" (UID: "c630f1e1-b605-4282-b136-647528bc86b4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.595015 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87853446-d88c-4d46-9fea-ff90af5617ba","Type":"ContainerStarted","Data":"a3f116584d27cd4ac1f92ca2ebf4ef11d519d12d09ab6730b129f1824cabea6f"} Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.612180 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c630f1e1-b605-4282-b136-647528bc86b4" (UID: "c630f1e1-b605-4282-b136-647528bc86b4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.617375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c630f1e1-b605-4282-b136-647528bc86b4-kube-api-access-dznxb" (OuterVolumeSpecName: "kube-api-access-dznxb") pod "c630f1e1-b605-4282-b136-647528bc86b4" (UID: "c630f1e1-b605-4282-b136-647528bc86b4"). InnerVolumeSpecName "kube-api-access-dznxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.630003 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.629985069 podStartE2EDuration="6.629985069s" podCreationTimestamp="2025-12-04 14:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:48.614938562 +0000 UTC m=+1345.683656787" watchObservedRunningTime="2025-12-04 14:19:48.629985069 +0000 UTC m=+1345.698703284" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662202 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 14:19:48 crc kubenswrapper[4715]: E1204 14:19:48.662638 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" containerName="init" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662656 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" containerName="init" Dec 04 14:19:48 crc kubenswrapper[4715]: E1204 14:19:48.662669 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="probe" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662675 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="probe" Dec 04 14:19:48 crc kubenswrapper[4715]: E1204 14:19:48.662704 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" containerName="dnsmasq-dns" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662710 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" containerName="dnsmasq-dns" Dec 04 14:19:48 crc kubenswrapper[4715]: E1204 14:19:48.662731 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="cinder-scheduler" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662737 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="cinder-scheduler" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662913 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="cinder-scheduler" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662932 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" containerName="dnsmasq-dns" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.662944 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c630f1e1-b605-4282-b136-647528bc86b4" containerName="probe" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.663602 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.667016 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.667272 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.667397 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-9b5x5" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.680557 4715 scope.go:117] "RemoveContainer" containerID="c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.695405 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dznxb\" (UniqueName: \"kubernetes.io/projected/c630f1e1-b605-4282-b136-647528bc86b4-kube-api-access-dznxb\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.695446 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.695518 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c630f1e1-b605-4282-b136-647528bc86b4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.699334 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.726943 4715 scope.go:117] "RemoveContainer" containerID="be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296" Dec 04 14:19:48 crc kubenswrapper[4715]: E1204 14:19:48.727610 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296\": container with ID starting with be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296 not found: ID does not exist" containerID="be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.727748 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296"} err="failed to get container status \"be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296\": rpc error: code = NotFound desc = could not find container \"be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296\": container with ID starting with be18de58a0fa531369ccd642e0cc8b05014a788f216f5dfc10e45e66f682c296 not found: ID does not exist" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.727871 4715 scope.go:117] "RemoveContainer" containerID="c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5" Dec 04 14:19:48 crc kubenswrapper[4715]: E1204 14:19:48.728506 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5\": container with ID starting with c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5 not found: ID does not exist" containerID="c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.728676 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5"} err="failed to get container status \"c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5\": rpc error: code = NotFound desc = could not find container \"c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5\": container with ID starting with c300cfb8d665a61d89085f38953c9c73f117214e0fae135667d776f0941c50c5 not found: ID does not exist" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.759913 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data" (OuterVolumeSpecName: "config-data") pod "c630f1e1-b605-4282-b136-647528bc86b4" (UID: "c630f1e1-b605-4282-b136-647528bc86b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.798102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-combined-ca-bundle\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.798278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-openstack-config\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.798301 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-openstack-config-secret\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.798472 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-584ng\" (UniqueName: \"kubernetes.io/projected/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-kube-api-access-584ng\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.798537 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c630f1e1-b605-4282-b136-647528bc86b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.899771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-584ng\" (UniqueName: \"kubernetes.io/projected/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-kube-api-access-584ng\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.899903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-combined-ca-bundle\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.900068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-openstack-config\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.900100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-openstack-config-secret\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.901581 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-openstack-config\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.904080 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-openstack-config-secret\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.905657 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-combined-ca-bundle\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.919869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-584ng\" (UniqueName: \"kubernetes.io/projected/454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72-kube-api-access-584ng\") pod \"openstackclient\" (UID: \"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72\") " pod="openstack/openstackclient" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.927723 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.934082 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.953247 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.954866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:19:48 crc kubenswrapper[4715]: I1204 14:19:48.957847 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.074556 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6816724d-3e76-4832-9e96-ab6ef342c354-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.074628 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.074774 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.074807 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52wn8\" (UniqueName: \"kubernetes.io/projected/6816724d-3e76-4832-9e96-ab6ef342c354-kube-api-access-52wn8\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.074838 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-config-data\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.074899 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-scripts\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.075570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.089228 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.176846 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.176888 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52wn8\" (UniqueName: \"kubernetes.io/projected/6816724d-3e76-4832-9e96-ab6ef342c354-kube-api-access-52wn8\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.176910 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-config-data\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.176958 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-scripts\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.176992 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6816724d-3e76-4832-9e96-ab6ef342c354-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.177020 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.178261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6816724d-3e76-4832-9e96-ab6ef342c354-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.184192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.184911 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-scripts\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.189530 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.193666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6816724d-3e76-4832-9e96-ab6ef342c354-config-data\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.199827 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52wn8\" (UniqueName: \"kubernetes.io/projected/6816724d-3e76-4832-9e96-ab6ef342c354-kube-api-access-52wn8\") pod \"cinder-scheduler-0\" (UID: \"6816724d-3e76-4832-9e96-ab6ef342c354\") " pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.213400 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a3bae3-f661-492c-bb10-cf636c135d06" path="/var/lib/kubelet/pods/92a3bae3-f661-492c-bb10-cf636c135d06/volumes" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.214404 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c630f1e1-b605-4282-b136-647528bc86b4" path="/var/lib/kubelet/pods/c630f1e1-b605-4282-b136-647528bc86b4/volumes" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.398825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.615746 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 14:19:49 crc kubenswrapper[4715]: I1204 14:19:49.750271 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 14:19:50 crc kubenswrapper[4715]: I1204 14:19:50.072688 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:19:50 crc kubenswrapper[4715]: I1204 14:19:50.655067 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:45628->10.217.0.146:8443: read: connection reset by peer" Dec 04 14:19:50 crc kubenswrapper[4715]: I1204 14:19:50.656128 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 14:19:50 crc kubenswrapper[4715]: I1204 14:19:50.781320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6816724d-3e76-4832-9e96-ab6ef342c354","Type":"ContainerStarted","Data":"1853aa8a17b3a3e4913fa164af133fa734e4d903a39897a737f532d51a861017"} Dec 04 14:19:50 crc kubenswrapper[4715]: I1204 14:19:50.800473 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72","Type":"ContainerStarted","Data":"2b87acd9df641a8904da6d052f3ec5274a4576d9cb8391d15f8f49b6e75b57ec"} Dec 04 14:19:51 crc kubenswrapper[4715]: I1204 14:19:51.813856 4715 generic.go:334] "Generic (PLEG): container finished" podID="88520da5-8fee-4a16-9396-2280c468c41a" containerID="f1297a46867ceb825461ae2c754dc4d5b3ba962c40fa5376fd6f94ccf7879eb0" exitCode=0 Dec 04 14:19:51 crc kubenswrapper[4715]: I1204 14:19:51.813943 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754dbdc69b-nwpzv" event={"ID":"88520da5-8fee-4a16-9396-2280c468c41a","Type":"ContainerDied","Data":"f1297a46867ceb825461ae2c754dc4d5b3ba962c40fa5376fd6f94ccf7879eb0"} Dec 04 14:19:51 crc kubenswrapper[4715]: I1204 14:19:51.817644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6816724d-3e76-4832-9e96-ab6ef342c354","Type":"ContainerStarted","Data":"7e33e43ce4e36b951b54ed8b8cc820717ce514eb7fde9f4e6b3fa4bc8a758d17"} Dec 04 14:19:52 crc kubenswrapper[4715]: I1204 14:19:52.833320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6816724d-3e76-4832-9e96-ab6ef342c354","Type":"ContainerStarted","Data":"4602a1b6e55ef6b38508178bc77ce6fee29d3e77637587dcfff94fe44fe582b7"} Dec 04 14:19:52 crc kubenswrapper[4715]: I1204 14:19:52.863663 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.863643074 podStartE2EDuration="4.863643074s" podCreationTimestamp="2025-12-04 14:19:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:19:52.853791257 +0000 UTC m=+1349.922509482" watchObservedRunningTime="2025-12-04 14:19:52.863643074 +0000 UTC m=+1349.932361289" Dec 04 14:19:54 crc kubenswrapper[4715]: I1204 14:19:54.399840 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 14:19:55 crc kubenswrapper[4715]: I1204 14:19:55.880897 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 14:19:58 crc kubenswrapper[4715]: I1204 14:19:58.805531 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.210312 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-64bbd97b75-ssj5n"] Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.213136 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.224867 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.226086 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.227315 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.229350 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-64bbd97b75-ssj5n"] Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.338998 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-config-data\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.339079 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-internal-tls-certs\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.339156 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861e33f9-3559-444f-8120-f439948fc774-run-httpd\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.339359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-combined-ca-bundle\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.339414 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p68xr\" (UniqueName: \"kubernetes.io/projected/861e33f9-3559-444f-8120-f439948fc774-kube-api-access-p68xr\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.339864 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861e33f9-3559-444f-8120-f439948fc774-log-httpd\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.342593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-public-tls-certs\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.342654 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/861e33f9-3559-444f-8120-f439948fc774-etc-swift\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.445783 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861e33f9-3559-444f-8120-f439948fc774-log-httpd\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.445846 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-public-tls-certs\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.445894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/861e33f9-3559-444f-8120-f439948fc774-etc-swift\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.445941 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-config-data\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.445970 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-internal-tls-certs\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.446005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861e33f9-3559-444f-8120-f439948fc774-run-httpd\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.446100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-combined-ca-bundle\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.446142 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p68xr\" (UniqueName: \"kubernetes.io/projected/861e33f9-3559-444f-8120-f439948fc774-kube-api-access-p68xr\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.446449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861e33f9-3559-444f-8120-f439948fc774-log-httpd\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.446481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861e33f9-3559-444f-8120-f439948fc774-run-httpd\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.454084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-internal-tls-certs\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.454204 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-public-tls-certs\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.454825 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-config-data\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.460457 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/861e33f9-3559-444f-8120-f439948fc774-etc-swift\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.466966 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p68xr\" (UniqueName: \"kubernetes.io/projected/861e33f9-3559-444f-8120-f439948fc774-kube-api-access-p68xr\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.471701 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861e33f9-3559-444f-8120-f439948fc774-combined-ca-bundle\") pod \"swift-proxy-64bbd97b75-ssj5n\" (UID: \"861e33f9-3559-444f-8120-f439948fc774\") " pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.549877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:19:59 crc kubenswrapper[4715]: I1204 14:19:59.966361 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 14:20:00 crc kubenswrapper[4715]: I1204 14:20:00.960655 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:00 crc kubenswrapper[4715]: I1204 14:20:00.961486 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-central-agent" containerID="cri-o://69e8a174021e90fc8244cb678755a51343fa78346c8ca9e1b8fc2cee10998bc9" gracePeriod=30 Dec 04 14:20:00 crc kubenswrapper[4715]: I1204 14:20:00.961710 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="sg-core" containerID="cri-o://321883fe0907f52c898143a19c0d0b867315c6dc23c0e187f48da043d04db650" gracePeriod=30 Dec 04 14:20:00 crc kubenswrapper[4715]: I1204 14:20:00.961777 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-notification-agent" containerID="cri-o://0d909ded27468e80eef0dad7f263f1a9c7f08e4812a144ca6407d3a5d0cd228c" gracePeriod=30 Dec 04 14:20:00 crc kubenswrapper[4715]: I1204 14:20:00.961710 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="proxy-httpd" containerID="cri-o://565247e23769c9075e305bf49b769a230bbb448dc04958fdb6dd922da3393a8c" gracePeriod=30 Dec 04 14:20:00 crc kubenswrapper[4715]: I1204 14:20:00.969207 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:20:01 crc kubenswrapper[4715]: I1204 14:20:01.973738 4715 generic.go:334] "Generic (PLEG): container finished" podID="268793f0-3a54-4a16-a059-7573491f9ef5" containerID="565247e23769c9075e305bf49b769a230bbb448dc04958fdb6dd922da3393a8c" exitCode=0 Dec 04 14:20:01 crc kubenswrapper[4715]: I1204 14:20:01.973780 4715 generic.go:334] "Generic (PLEG): container finished" podID="268793f0-3a54-4a16-a059-7573491f9ef5" containerID="321883fe0907f52c898143a19c0d0b867315c6dc23c0e187f48da043d04db650" exitCode=2 Dec 04 14:20:01 crc kubenswrapper[4715]: I1204 14:20:01.973794 4715 generic.go:334] "Generic (PLEG): container finished" podID="268793f0-3a54-4a16-a059-7573491f9ef5" containerID="69e8a174021e90fc8244cb678755a51343fa78346c8ca9e1b8fc2cee10998bc9" exitCode=0 Dec 04 14:20:01 crc kubenswrapper[4715]: I1204 14:20:01.973819 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerDied","Data":"565247e23769c9075e305bf49b769a230bbb448dc04958fdb6dd922da3393a8c"} Dec 04 14:20:01 crc kubenswrapper[4715]: I1204 14:20:01.973850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerDied","Data":"321883fe0907f52c898143a19c0d0b867315c6dc23c0e187f48da043d04db650"} Dec 04 14:20:01 crc kubenswrapper[4715]: I1204 14:20:01.973863 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerDied","Data":"69e8a174021e90fc8244cb678755a51343fa78346c8ca9e1b8fc2cee10998bc9"} Dec 04 14:20:02 crc kubenswrapper[4715]: I1204 14:20:02.991243 4715 generic.go:334] "Generic (PLEG): container finished" podID="268793f0-3a54-4a16-a059-7573491f9ef5" containerID="0d909ded27468e80eef0dad7f263f1a9c7f08e4812a144ca6407d3a5d0cd228c" exitCode=0 Dec 04 14:20:02 crc kubenswrapper[4715]: I1204 14:20:02.991315 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerDied","Data":"0d909ded27468e80eef0dad7f263f1a9c7f08e4812a144ca6407d3a5d0cd228c"} Dec 04 14:20:04 crc kubenswrapper[4715]: I1204 14:20:04.512423 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:20:04 crc kubenswrapper[4715]: I1204 14:20:04.768413 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:20:04 crc kubenswrapper[4715]: I1204 14:20:04.768988 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-log" containerID="cri-o://c5b5d7b5a1ff4e61fcd49d9a1d37cf050a7d83db8680451c80b6e2e5269ec17e" gracePeriod=30 Dec 04 14:20:04 crc kubenswrapper[4715]: I1204 14:20:04.769162 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-httpd" containerID="cri-o://74c00410246b6ff5e0221ad7de293d016e8bcf25cd908b648f975a4bc1b1aeba" gracePeriod=30 Dec 04 14:20:05 crc kubenswrapper[4715]: I1204 14:20:05.029530 4715 generic.go:334] "Generic (PLEG): container finished" podID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerID="c5b5d7b5a1ff4e61fcd49d9a1d37cf050a7d83db8680451c80b6e2e5269ec17e" exitCode=143 Dec 04 14:20:05 crc kubenswrapper[4715]: I1204 14:20:05.029591 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"edf7d465-56e6-4185-8ac7-bdbff1e10866","Type":"ContainerDied","Data":"c5b5d7b5a1ff4e61fcd49d9a1d37cf050a7d83db8680451c80b6e2e5269ec17e"} Dec 04 14:20:06 crc kubenswrapper[4715]: E1204 14:20:06.385556 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Dec 04 14:20:06 crc kubenswrapper[4715]: E1204 14:20:06.386702 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bdh58dh5cdhcbh5c4h9chb7hf8h694h54bhbh6fhc5h94h74h58ch6fh5b5hc8h64bh5c6h5bch88h656h58fh697h5cdhc7h4h5bch5d9h94q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-584ng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:20:06 crc kubenswrapper[4715]: E1204 14:20:06.388157 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72" Dec 04 14:20:06 crc kubenswrapper[4715]: I1204 14:20:06.905657 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4x84z"] Dec 04 14:20:06 crc kubenswrapper[4715]: I1204 14:20:06.916528 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:06 crc kubenswrapper[4715]: I1204 14:20:06.934559 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4x84z"] Dec 04 14:20:06 crc kubenswrapper[4715]: I1204 14:20:06.946641 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005133 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-ccnsf"] Dec 04 14:20:07 crc kubenswrapper[4715]: E1204 14:20:07.005573 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-notification-agent" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005591 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-notification-agent" Dec 04 14:20:07 crc kubenswrapper[4715]: E1204 14:20:07.005610 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="sg-core" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005616 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="sg-core" Dec 04 14:20:07 crc kubenswrapper[4715]: E1204 14:20:07.005663 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-central-agent" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005671 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-central-agent" Dec 04 14:20:07 crc kubenswrapper[4715]: E1204 14:20:07.005683 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="proxy-httpd" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005690 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="proxy-httpd" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005886 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-notification-agent" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005903 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="sg-core" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005914 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="proxy-httpd" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.005928 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" containerName="ceilometer-central-agent" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.006661 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.040541 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2139a8dc-19bb-453f-a293-02b7f5944810-operator-scripts\") pod \"nova-api-db-create-4x84z\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.040602 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b9dc\" (UniqueName: \"kubernetes.io/projected/2139a8dc-19bb-453f-a293-02b7f5944810-kube-api-access-6b9dc\") pod \"nova-api-db-create-4x84z\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.054307 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ccnsf"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.096010 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.096311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"268793f0-3a54-4a16-a059-7573491f9ef5","Type":"ContainerDied","Data":"9a27e84cb32ef373108f2f05b949121de6303d1f04893ed036248d3604ec1676"} Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.096353 4715 scope.go:117] "RemoveContainer" containerID="565247e23769c9075e305bf49b769a230bbb448dc04958fdb6dd922da3393a8c" Dec 04 14:20:07 crc kubenswrapper[4715]: E1204 14:20:07.099187 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.130602 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wbgxz"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.132140 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.140260 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-01e0-account-create-update-h84ff"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.159806 4715 scope.go:117] "RemoveContainer" containerID="321883fe0907f52c898143a19c0d0b867315c6dc23c0e187f48da043d04db650" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.160561 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.163890 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164154 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-scripts\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-config-data\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164370 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-combined-ca-bundle\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164486 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-run-httpd\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164525 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-sg-core-conf-yaml\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164612 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wwbt\" (UniqueName: \"kubernetes.io/projected/268793f0-3a54-4a16-a059-7573491f9ef5-kube-api-access-9wwbt\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.164742 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-log-httpd\") pod \"268793f0-3a54-4a16-a059-7573491f9ef5\" (UID: \"268793f0-3a54-4a16-a059-7573491f9ef5\") " Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.165859 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff544d83-9fa6-444c-bb1e-212318e8dc6a-operator-scripts\") pod \"nova-cell1-db-create-wbgxz\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.166458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpggt\" (UniqueName: \"kubernetes.io/projected/d49248bc-1dc1-45c2-b0a5-6886331d3c54-kube-api-access-mpggt\") pod \"nova-cell0-db-create-ccnsf\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.166670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d49248bc-1dc1-45c2-b0a5-6886331d3c54-operator-scripts\") pod \"nova-cell0-db-create-ccnsf\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.166734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l9x2\" (UniqueName: \"kubernetes.io/projected/584d122e-147a-48a3-993e-6ecd0ee9208a-kube-api-access-6l9x2\") pod \"nova-api-01e0-account-create-update-h84ff\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.166796 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d122e-147a-48a3-993e-6ecd0ee9208a-operator-scripts\") pod \"nova-api-01e0-account-create-update-h84ff\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.166843 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfvqf\" (UniqueName: \"kubernetes.io/projected/ff544d83-9fa6-444c-bb1e-212318e8dc6a-kube-api-access-bfvqf\") pod \"nova-cell1-db-create-wbgxz\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.167247 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2139a8dc-19bb-453f-a293-02b7f5944810-operator-scripts\") pod \"nova-api-db-create-4x84z\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.168729 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b9dc\" (UniqueName: \"kubernetes.io/projected/2139a8dc-19bb-453f-a293-02b7f5944810-kube-api-access-6b9dc\") pod \"nova-api-db-create-4x84z\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.176574 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-scripts" (OuterVolumeSpecName: "scripts") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.177666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2139a8dc-19bb-453f-a293-02b7f5944810-operator-scripts\") pod \"nova-api-db-create-4x84z\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.181455 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.181690 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.199054 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268793f0-3a54-4a16-a059-7573491f9ef5-kube-api-access-9wwbt" (OuterVolumeSpecName: "kube-api-access-9wwbt") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "kube-api-access-9wwbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.276560 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpggt\" (UniqueName: \"kubernetes.io/projected/d49248bc-1dc1-45c2-b0a5-6886331d3c54-kube-api-access-mpggt\") pod \"nova-cell0-db-create-ccnsf\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.276674 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d49248bc-1dc1-45c2-b0a5-6886331d3c54-operator-scripts\") pod \"nova-cell0-db-create-ccnsf\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.276726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l9x2\" (UniqueName: \"kubernetes.io/projected/584d122e-147a-48a3-993e-6ecd0ee9208a-kube-api-access-6l9x2\") pod \"nova-api-01e0-account-create-update-h84ff\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.276842 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d122e-147a-48a3-993e-6ecd0ee9208a-operator-scripts\") pod \"nova-api-01e0-account-create-update-h84ff\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.276894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfvqf\" (UniqueName: \"kubernetes.io/projected/ff544d83-9fa6-444c-bb1e-212318e8dc6a-kube-api-access-bfvqf\") pod \"nova-cell1-db-create-wbgxz\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.277105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff544d83-9fa6-444c-bb1e-212318e8dc6a-operator-scripts\") pod \"nova-cell1-db-create-wbgxz\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.279210 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.288792 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wwbt\" (UniqueName: \"kubernetes.io/projected/268793f0-3a54-4a16-a059-7573491f9ef5-kube-api-access-9wwbt\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.288843 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/268793f0-3a54-4a16-a059-7573491f9ef5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.288858 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.290879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff544d83-9fa6-444c-bb1e-212318e8dc6a-operator-scripts\") pod \"nova-cell1-db-create-wbgxz\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.293958 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d49248bc-1dc1-45c2-b0a5-6886331d3c54-operator-scripts\") pod \"nova-cell0-db-create-ccnsf\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.307998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d122e-147a-48a3-993e-6ecd0ee9208a-operator-scripts\") pod \"nova-api-01e0-account-create-update-h84ff\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.312708 4715 scope.go:117] "RemoveContainer" containerID="0d909ded27468e80eef0dad7f263f1a9c7f08e4812a144ca6407d3a5d0cd228c" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.340754 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.346322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b9dc\" (UniqueName: \"kubernetes.io/projected/2139a8dc-19bb-453f-a293-02b7f5944810-kube-api-access-6b9dc\") pod \"nova-api-db-create-4x84z\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.346546 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfvqf\" (UniqueName: \"kubernetes.io/projected/ff544d83-9fa6-444c-bb1e-212318e8dc6a-kube-api-access-bfvqf\") pod \"nova-cell1-db-create-wbgxz\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.356527 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l9x2\" (UniqueName: \"kubernetes.io/projected/584d122e-147a-48a3-993e-6ecd0ee9208a-kube-api-access-6l9x2\") pod \"nova-api-01e0-account-create-update-h84ff\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.361590 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpggt\" (UniqueName: \"kubernetes.io/projected/d49248bc-1dc1-45c2-b0a5-6886331d3c54-kube-api-access-mpggt\") pod \"nova-cell0-db-create-ccnsf\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.391937 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.464580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-config-data" (OuterVolumeSpecName: "config-data") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.485813 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "268793f0-3a54-4a16-a059-7573491f9ef5" (UID: "268793f0-3a54-4a16-a059-7573491f9ef5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.496105 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.496145 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268793f0-3a54-4a16-a059-7573491f9ef5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.546642 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wbgxz"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.546995 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-01e0-account-create-update-h84ff"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.547023 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-59d3-account-create-update-xm5ts"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.548537 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-59d3-account-create-update-xm5ts"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.548563 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-42a0-account-create-update-jhj8z"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.548779 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.549616 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-42a0-account-create-update-jhj8z"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.549720 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.556234 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.567571 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.567942 4715 scope.go:117] "RemoveContainer" containerID="69e8a174021e90fc8244cb678755a51343fa78346c8ca9e1b8fc2cee10998bc9" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.571765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.600652 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-64bbd97b75-ssj5n"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.611306 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.611919 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.636434 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.646423 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.646705 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-log" containerID="cri-o://34ebffad1ccd13f9dd6113bd6daea6476e89742be07724b139fd53e1c3592d0b" gracePeriod=30 Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.646865 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-httpd" containerID="cri-o://4e483519ce0acbca915c0fd7e57849d6f3637c62329e5d4be0e6008cdddc6593" gracePeriod=30 Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.700463 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l5pr\" (UniqueName: \"kubernetes.io/projected/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-kube-api-access-8l5pr\") pod \"nova-cell0-59d3-account-create-update-xm5ts\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.700539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28t8n\" (UniqueName: \"kubernetes.io/projected/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-kube-api-access-28t8n\") pod \"nova-cell1-42a0-account-create-update-jhj8z\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.700574 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-operator-scripts\") pod \"nova-cell1-42a0-account-create-update-jhj8z\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.700652 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-operator-scripts\") pod \"nova-cell0-59d3-account-create-update-xm5ts\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.796282 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.803508 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l5pr\" (UniqueName: \"kubernetes.io/projected/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-kube-api-access-8l5pr\") pod \"nova-cell0-59d3-account-create-update-xm5ts\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.803972 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28t8n\" (UniqueName: \"kubernetes.io/projected/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-kube-api-access-28t8n\") pod \"nova-cell1-42a0-account-create-update-jhj8z\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.804365 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-operator-scripts\") pod \"nova-cell1-42a0-account-create-update-jhj8z\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.804560 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-operator-scripts\") pod \"nova-cell0-59d3-account-create-update-xm5ts\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.806462 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-operator-scripts\") pod \"nova-cell0-59d3-account-create-update-xm5ts\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.808559 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-operator-scripts\") pod \"nova-cell1-42a0-account-create-update-jhj8z\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.834760 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28t8n\" (UniqueName: \"kubernetes.io/projected/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-kube-api-access-28t8n\") pod \"nova-cell1-42a0-account-create-update-jhj8z\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.844736 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l5pr\" (UniqueName: \"kubernetes.io/projected/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-kube-api-access-8l5pr\") pod \"nova-cell0-59d3-account-create-update-xm5ts\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.851446 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.862186 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.871051 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.871448 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.873982 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.874084 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.876685 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:07 crc kubenswrapper[4715]: I1204 14:20:07.890400 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010343 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-config-data\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-log-httpd\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010505 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgmdh\" (UniqueName: \"kubernetes.io/projected/b80b13f4-50d2-4586-a7c5-5cf18273bb48-kube-api-access-tgmdh\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010540 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-scripts\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.010580 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-run-httpd\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-config-data\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113766 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-log-httpd\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgmdh\" (UniqueName: \"kubernetes.io/projected/b80b13f4-50d2-4586-a7c5-5cf18273bb48-kube-api-access-tgmdh\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113931 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-scripts\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.113996 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-run-httpd\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.114947 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-run-httpd\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.115682 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-log-httpd\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.122192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-config-data\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.124221 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-scripts\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.124410 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.135366 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.137127 4715 generic.go:334] "Generic (PLEG): container finished" podID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerID="34ebffad1ccd13f9dd6113bd6daea6476e89742be07724b139fd53e1c3592d0b" exitCode=143 Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.137463 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d95418a-f1f5-4676-903e-67b2fee4417c","Type":"ContainerDied","Data":"34ebffad1ccd13f9dd6113bd6daea6476e89742be07724b139fd53e1c3592d0b"} Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.138603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgmdh\" (UniqueName: \"kubernetes.io/projected/b80b13f4-50d2-4586-a7c5-5cf18273bb48-kube-api-access-tgmdh\") pod \"ceilometer-0\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.154277 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64bbd97b75-ssj5n" event={"ID":"861e33f9-3559-444f-8120-f439948fc774","Type":"ContainerStarted","Data":"8b156313618c0bc642a3894bd79974194c0b740703d71cd246f56fc5aec7d485"} Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.239205 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.270826 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4x84z"] Dec 04 14:20:08 crc kubenswrapper[4715]: W1204 14:20:08.291710 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2139a8dc_19bb_453f_a293_02b7f5944810.slice/crio-6d407af0a818bb2ff183a83537995f6224ba2aa73cff1b593dbd83b91e4016ed WatchSource:0}: Error finding container 6d407af0a818bb2ff183a83537995f6224ba2aa73cff1b593dbd83b91e4016ed: Status 404 returned error can't find the container with id 6d407af0a818bb2ff183a83537995f6224ba2aa73cff1b593dbd83b91e4016ed Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.437654 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-01e0-account-create-update-h84ff"] Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.576884 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ccnsf"] Dec 04 14:20:08 crc kubenswrapper[4715]: W1204 14:20:08.586811 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd49248bc_1dc1_45c2_b0a5_6886331d3c54.slice/crio-b6694ab2f3e150755c7ec02cd9a2baa9b81151b91623df4e4825c241e27d2e33 WatchSource:0}: Error finding container b6694ab2f3e150755c7ec02cd9a2baa9b81151b91623df4e4825c241e27d2e33: Status 404 returned error can't find the container with id b6694ab2f3e150755c7ec02cd9a2baa9b81151b91623df4e4825c241e27d2e33 Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.689515 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-42a0-account-create-update-jhj8z"] Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.699774 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wbgxz"] Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.770739 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 14:20:08 crc kubenswrapper[4715]: I1204 14:20:08.965545 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-59d3-account-create-update-xm5ts"] Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.005868 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.172433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ccnsf" event={"ID":"d49248bc-1dc1-45c2-b0a5-6886331d3c54","Type":"ContainerStarted","Data":"b6694ab2f3e150755c7ec02cd9a2baa9b81151b91623df4e4825c241e27d2e33"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.195871 4715 generic.go:334] "Generic (PLEG): container finished" podID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerID="74c00410246b6ff5e0221ad7de293d016e8bcf25cd908b648f975a4bc1b1aeba" exitCode=0 Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.199933 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="268793f0-3a54-4a16-a059-7573491f9ef5" path="/var/lib/kubelet/pods/268793f0-3a54-4a16-a059-7573491f9ef5/volumes" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.200928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"edf7d465-56e6-4185-8ac7-bdbff1e10866","Type":"ContainerDied","Data":"74c00410246b6ff5e0221ad7de293d016e8bcf25cd908b648f975a4bc1b1aeba"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.203935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" event={"ID":"13239d2e-d649-4a1c-aafc-fbdc135d7c2f","Type":"ContainerStarted","Data":"5bc002f7d3eea49a690646bdec63fcd1119f939c885700f7f7a2f0bc0c1ef5eb"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.210957 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbgxz" event={"ID":"ff544d83-9fa6-444c-bb1e-212318e8dc6a","Type":"ContainerStarted","Data":"09c546d6e71b7e440f67c0fda56877cb77057511626eee5633e6f11f359fbaea"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.216453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4x84z" event={"ID":"2139a8dc-19bb-453f-a293-02b7f5944810","Type":"ContainerStarted","Data":"6d407af0a818bb2ff183a83537995f6224ba2aa73cff1b593dbd83b91e4016ed"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.220933 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerStarted","Data":"0d9f99f4811808b653a6e6241de5333a069afe6853e3decf84f74e97c01bbff8"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.257484 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-01e0-account-create-update-h84ff" event={"ID":"584d122e-147a-48a3-993e-6ecd0ee9208a","Type":"ContainerStarted","Data":"2b02a262ff9daf6600450234fbb61c2d89e08f56ced9026a88cbcb431d97f1c9"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.254563 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-4x84z" podStartSLOduration=3.254504913 podStartE2EDuration="3.254504913s" podCreationTimestamp="2025-12-04 14:20:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:20:09.253302571 +0000 UTC m=+1366.322020796" watchObservedRunningTime="2025-12-04 14:20:09.254504913 +0000 UTC m=+1366.323223138" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.271186 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" event={"ID":"f1e7bb9b-d971-4641-b02d-0e17b0afe36f","Type":"ContainerStarted","Data":"12e6a535d42dbbb072a58a0f18a90fdbab8f19dda59cf4f6600b9f3073526801"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.277852 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.281782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64bbd97b75-ssj5n" event={"ID":"861e33f9-3559-444f-8120-f439948fc774","Type":"ContainerStarted","Data":"5ebbec55117f690d49c42ab2b5efd0ee53877bc04c1020de7dda7ce57e9c9ee8"} Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.363671 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-logs\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.363801 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-httpd-run\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.363872 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.363910 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-scripts\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.363980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dsh8\" (UniqueName: \"kubernetes.io/projected/edf7d465-56e6-4185-8ac7-bdbff1e10866-kube-api-access-7dsh8\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.364065 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-public-tls-certs\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.364111 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-config-data\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.364172 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-combined-ca-bundle\") pod \"edf7d465-56e6-4185-8ac7-bdbff1e10866\" (UID: \"edf7d465-56e6-4185-8ac7-bdbff1e10866\") " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.369154 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-logs" (OuterVolumeSpecName: "logs") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.377078 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.379590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.393275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf7d465-56e6-4185-8ac7-bdbff1e10866-kube-api-access-7dsh8" (OuterVolumeSpecName: "kube-api-access-7dsh8") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "kube-api-access-7dsh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.403443 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-scripts" (OuterVolumeSpecName: "scripts") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.467559 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.467606 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edf7d465-56e6-4185-8ac7-bdbff1e10866-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.467633 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.467649 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.467661 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dsh8\" (UniqueName: \"kubernetes.io/projected/edf7d465-56e6-4185-8ac7-bdbff1e10866-kube-api-access-7dsh8\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.646344 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.662698 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.668723 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-config-data" (OuterVolumeSpecName: "config-data") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.672281 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.672310 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.672321 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.720179 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "edf7d465-56e6-4185-8ac7-bdbff1e10866" (UID: "edf7d465-56e6-4185-8ac7-bdbff1e10866"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.774421 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf7d465-56e6-4185-8ac7-bdbff1e10866-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:09 crc kubenswrapper[4715]: I1204 14:20:09.976110 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.296315 4715 generic.go:334] "Generic (PLEG): container finished" podID="d49248bc-1dc1-45c2-b0a5-6886331d3c54" containerID="f1b186f73b402152779c2509ccc329465c4fd3c1d6c039a8d5ccfa24e80628c9" exitCode=0 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.296416 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ccnsf" event={"ID":"d49248bc-1dc1-45c2-b0a5-6886331d3c54","Type":"ContainerDied","Data":"f1b186f73b402152779c2509ccc329465c4fd3c1d6c039a8d5ccfa24e80628c9"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.300292 4715 generic.go:334] "Generic (PLEG): container finished" podID="13239d2e-d649-4a1c-aafc-fbdc135d7c2f" containerID="673c9251cbe5e8bf0d54b5ba990a3a137dac691cd8675e5c827f151f4d8a2db9" exitCode=0 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.300361 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" event={"ID":"13239d2e-d649-4a1c-aafc-fbdc135d7c2f","Type":"ContainerDied","Data":"673c9251cbe5e8bf0d54b5ba990a3a137dac691cd8675e5c827f151f4d8a2db9"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.303630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbgxz" event={"ID":"ff544d83-9fa6-444c-bb1e-212318e8dc6a","Type":"ContainerStarted","Data":"ba408ab6eeb9b2334ad7a5ffb2d27ea1e1b2d5007dd845cf518b54e9879253a0"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.305970 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139a8dc-19bb-453f-a293-02b7f5944810" containerID="196ac5af78f8eec8cc7897c6da207f2c04d7f87ad6219743bfafb851eedbc235" exitCode=0 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.306127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4x84z" event={"ID":"2139a8dc-19bb-453f-a293-02b7f5944810","Type":"ContainerDied","Data":"196ac5af78f8eec8cc7897c6da207f2c04d7f87ad6219743bfafb851eedbc235"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.309876 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64bbd97b75-ssj5n" event={"ID":"861e33f9-3559-444f-8120-f439948fc774","Type":"ContainerStarted","Data":"a716dfdc84bc4437739877c47d7813a73323ab6848074f0bc3c7c9b169a9be5c"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.310837 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.310870 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.312370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerStarted","Data":"1e630ce7abae93726166f066bace8ceed01d5e310f1daebdb2ce9780eaeebc7a"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.313461 4715 generic.go:334] "Generic (PLEG): container finished" podID="584d122e-147a-48a3-993e-6ecd0ee9208a" containerID="d98000028662d7255e40983fd272c975c47a90a63b1956dca5731c2fc5ec5540" exitCode=0 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.313506 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-01e0-account-create-update-h84ff" event={"ID":"584d122e-147a-48a3-993e-6ecd0ee9208a","Type":"ContainerDied","Data":"d98000028662d7255e40983fd272c975c47a90a63b1956dca5731c2fc5ec5540"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.315611 4715 generic.go:334] "Generic (PLEG): container finished" podID="f1e7bb9b-d971-4641-b02d-0e17b0afe36f" containerID="bd271d86020cadb182cd20b55e5f1a3660bb60b7b70733a1ab48dd7fa5d04273" exitCode=0 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.315657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" event={"ID":"f1e7bb9b-d971-4641-b02d-0e17b0afe36f","Type":"ContainerDied","Data":"bd271d86020cadb182cd20b55e5f1a3660bb60b7b70733a1ab48dd7fa5d04273"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.317699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"edf7d465-56e6-4185-8ac7-bdbff1e10866","Type":"ContainerDied","Data":"6e497f9693dac055f4d3242f82848f44442908a01e68eea7344e1815df06829b"} Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.317741 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.317764 4715 scope.go:117] "RemoveContainer" containerID="74c00410246b6ff5e0221ad7de293d016e8bcf25cd908b648f975a4bc1b1aeba" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.338479 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-wbgxz" podStartSLOduration=3.33845629 podStartE2EDuration="3.33845629s" podCreationTimestamp="2025-12-04 14:20:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:20:10.325638782 +0000 UTC m=+1367.394357007" watchObservedRunningTime="2025-12-04 14:20:10.33845629 +0000 UTC m=+1367.407174515" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.350114 4715 scope.go:117] "RemoveContainer" containerID="c5b5d7b5a1ff4e61fcd49d9a1d37cf050a7d83db8680451c80b6e2e5269ec17e" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.418948 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.443190 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.450863 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-64bbd97b75-ssj5n" podStartSLOduration=11.450834943 podStartE2EDuration="11.450834943s" podCreationTimestamp="2025-12-04 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:20:10.410209473 +0000 UTC m=+1367.478927708" watchObservedRunningTime="2025-12-04 14:20:10.450834943 +0000 UTC m=+1367.519553168" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.479679 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:20:10 crc kubenswrapper[4715]: E1204 14:20:10.480208 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-httpd" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.480226 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-httpd" Dec 04 14:20:10 crc kubenswrapper[4715]: E1204 14:20:10.480261 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-log" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.480271 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-log" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.480509 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-httpd" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.480530 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" containerName="glance-log" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.481737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.485457 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.485764 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.499931 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.590500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92c1ce5a-911d-4567-8d23-11777a6df443-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.590834 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-config-data\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.590899 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92c1ce5a-911d-4567-8d23-11777a6df443-logs\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.591001 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-scripts\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.591063 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldbr8\" (UniqueName: \"kubernetes.io/projected/92c1ce5a-911d-4567-8d23-11777a6df443-kube-api-access-ldbr8\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.591119 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.591246 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.591292 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.611278 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-58655b5d9c-m5lkw" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.696983 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92c1ce5a-911d-4567-8d23-11777a6df443-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.697198 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-config-data\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.697370 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92c1ce5a-911d-4567-8d23-11777a6df443-logs\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.697560 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-scripts\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.697768 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldbr8\" (UniqueName: \"kubernetes.io/projected/92c1ce5a-911d-4567-8d23-11777a6df443-kube-api-access-ldbr8\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.707632 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d7bd77bfb-nlk9m"] Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.707978 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d7bd77bfb-nlk9m" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-api" containerID="cri-o://dd7b9e026fbce19f88bd2063857eba783cb3f3f66bde675059e0b4d466564778" gracePeriod=30 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.708333 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d7bd77bfb-nlk9m" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-httpd" containerID="cri-o://0f42ca9a061f45d393889a4cfa54ca235d7648375ad5d892f8dc431de02d3d95" gracePeriod=30 Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.711151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92c1ce5a-911d-4567-8d23-11777a6df443-logs\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.711466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92c1ce5a-911d-4567-8d23-11777a6df443-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.713208 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.713524 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.713676 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.717210 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.719501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.729360 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-config-data\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.736708 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldbr8\" (UniqueName: \"kubernetes.io/projected/92c1ce5a-911d-4567-8d23-11777a6df443-kube-api-access-ldbr8\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.747157 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-scripts\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.748885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c1ce5a-911d-4567-8d23-11777a6df443-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:10 crc kubenswrapper[4715]: I1204 14:20:10.863648 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"92c1ce5a-911d-4567-8d23-11777a6df443\") " pod="openstack/glance-default-external-api-0" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.101352 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.208442 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf7d465-56e6-4185-8ac7-bdbff1e10866" path="/var/lib/kubelet/pods/edf7d465-56e6-4185-8ac7-bdbff1e10866/volumes" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.398343 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerStarted","Data":"b86ff9faafd6567d9e9147f726cf05b3e0ce4097006bd21594161baa66c853f8"} Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.423006 4715 generic.go:334] "Generic (PLEG): container finished" podID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerID="0f42ca9a061f45d393889a4cfa54ca235d7648375ad5d892f8dc431de02d3d95" exitCode=0 Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.423144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7bd77bfb-nlk9m" event={"ID":"d49eba49-c182-4ec6-9a61-d14cddf541fd","Type":"ContainerDied","Data":"0f42ca9a061f45d393889a4cfa54ca235d7648375ad5d892f8dc431de02d3d95"} Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.442325 4715 generic.go:334] "Generic (PLEG): container finished" podID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerID="4e483519ce0acbca915c0fd7e57849d6f3637c62329e5d4be0e6008cdddc6593" exitCode=0 Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.442437 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d95418a-f1f5-4676-903e-67b2fee4417c","Type":"ContainerDied","Data":"4e483519ce0acbca915c0fd7e57849d6f3637c62329e5d4be0e6008cdddc6593"} Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.468325 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff544d83-9fa6-444c-bb1e-212318e8dc6a" containerID="ba408ab6eeb9b2334ad7a5ffb2d27ea1e1b2d5007dd845cf518b54e9879253a0" exitCode=0 Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.468881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbgxz" event={"ID":"ff544d83-9fa6-444c-bb1e-212318e8dc6a","Type":"ContainerDied","Data":"ba408ab6eeb9b2334ad7a5ffb2d27ea1e1b2d5007dd845cf518b54e9879253a0"} Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.699541 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.841754 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842134 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-internal-tls-certs\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842200 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-scripts\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842225 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-combined-ca-bundle\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842330 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-httpd-run\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842417 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp5x6\" (UniqueName: \"kubernetes.io/projected/7d95418a-f1f5-4676-903e-67b2fee4417c-kube-api-access-hp5x6\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842467 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-logs\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.842498 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-config-data\") pod \"7d95418a-f1f5-4676-903e-67b2fee4417c\" (UID: \"7d95418a-f1f5-4676-903e-67b2fee4417c\") " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.843842 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.844129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-logs" (OuterVolumeSpecName: "logs") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.851897 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.858286 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d95418a-f1f5-4676-903e-67b2fee4417c-kube-api-access-hp5x6" (OuterVolumeSpecName: "kube-api-access-hp5x6") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "kube-api-access-hp5x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.860257 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-scripts" (OuterVolumeSpecName: "scripts") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.898224 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.962253 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.962287 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.962297 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.962307 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp5x6\" (UniqueName: \"kubernetes.io/projected/7d95418a-f1f5-4676-903e-67b2fee4417c-kube-api-access-hp5x6\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.962317 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d95418a-f1f5-4676-903e-67b2fee4417c-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.962338 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 04 14:20:11 crc kubenswrapper[4715]: I1204 14:20:11.994230 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.011254 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.018266 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.018619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-config-data" (OuterVolumeSpecName: "config-data") pod "7d95418a-f1f5-4676-903e-67b2fee4417c" (UID: "7d95418a-f1f5-4676-903e-67b2fee4417c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.071315 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.071572 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d95418a-f1f5-4676-903e-67b2fee4417c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.071669 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.302731 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.303667 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.357989 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.381246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d49248bc-1dc1-45c2-b0a5-6886331d3c54-operator-scripts\") pod \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.381907 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-operator-scripts\") pod \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.382185 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l5pr\" (UniqueName: \"kubernetes.io/projected/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-kube-api-access-8l5pr\") pod \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\" (UID: \"f1e7bb9b-d971-4641-b02d-0e17b0afe36f\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.383933 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpggt\" (UniqueName: \"kubernetes.io/projected/d49248bc-1dc1-45c2-b0a5-6886331d3c54-kube-api-access-mpggt\") pod \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\" (UID: \"d49248bc-1dc1-45c2-b0a5-6886331d3c54\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.384568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49248bc-1dc1-45c2-b0a5-6886331d3c54-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d49248bc-1dc1-45c2-b0a5-6886331d3c54" (UID: "d49248bc-1dc1-45c2-b0a5-6886331d3c54"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.385271 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d49248bc-1dc1-45c2-b0a5-6886331d3c54-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.391663 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-kube-api-access-8l5pr" (OuterVolumeSpecName: "kube-api-access-8l5pr") pod "f1e7bb9b-d971-4641-b02d-0e17b0afe36f" (UID: "f1e7bb9b-d971-4641-b02d-0e17b0afe36f"). InnerVolumeSpecName "kube-api-access-8l5pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.392512 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f1e7bb9b-d971-4641-b02d-0e17b0afe36f" (UID: "f1e7bb9b-d971-4641-b02d-0e17b0afe36f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.393299 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49248bc-1dc1-45c2-b0a5-6886331d3c54-kube-api-access-mpggt" (OuterVolumeSpecName: "kube-api-access-mpggt") pod "d49248bc-1dc1-45c2-b0a5-6886331d3c54" (UID: "d49248bc-1dc1-45c2-b0a5-6886331d3c54"). InnerVolumeSpecName "kube-api-access-mpggt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.459640 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.461244 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.473287 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.490080 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.490109 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l5pr\" (UniqueName: \"kubernetes.io/projected/f1e7bb9b-d971-4641-b02d-0e17b0afe36f-kube-api-access-8l5pr\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.490119 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpggt\" (UniqueName: \"kubernetes.io/projected/d49248bc-1dc1-45c2-b0a5-6886331d3c54-kube-api-access-mpggt\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.497320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ccnsf" event={"ID":"d49248bc-1dc1-45c2-b0a5-6886331d3c54","Type":"ContainerDied","Data":"b6694ab2f3e150755c7ec02cd9a2baa9b81151b91623df4e4825c241e27d2e33"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.497361 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6694ab2f3e150755c7ec02cd9a2baa9b81151b91623df4e4825c241e27d2e33" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.497429 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ccnsf" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.526070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" event={"ID":"f1e7bb9b-d971-4641-b02d-0e17b0afe36f","Type":"ContainerDied","Data":"12e6a535d42dbbb072a58a0f18a90fdbab8f19dda59cf4f6600b9f3073526801"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.526111 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12e6a535d42dbbb072a58a0f18a90fdbab8f19dda59cf4f6600b9f3073526801" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.526162 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-59d3-account-create-update-xm5ts" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.540160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d95418a-f1f5-4676-903e-67b2fee4417c","Type":"ContainerDied","Data":"e3283855985c227b217b7a79ef3db5c1b02913560cf0d652b37702839d2a5481"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.540207 4715 scope.go:117] "RemoveContainer" containerID="4e483519ce0acbca915c0fd7e57849d6f3637c62329e5d4be0e6008cdddc6593" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.540358 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.576271 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" event={"ID":"13239d2e-d649-4a1c-aafc-fbdc135d7c2f","Type":"ContainerDied","Data":"5bc002f7d3eea49a690646bdec63fcd1119f939c885700f7f7a2f0bc0c1ef5eb"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.576312 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bc002f7d3eea49a690646bdec63fcd1119f939c885700f7f7a2f0bc0c1ef5eb" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.576382 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-42a0-account-create-update-jhj8z" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.594692 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d122e-147a-48a3-993e-6ecd0ee9208a-operator-scripts\") pod \"584d122e-147a-48a3-993e-6ecd0ee9208a\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.594745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2139a8dc-19bb-453f-a293-02b7f5944810-operator-scripts\") pod \"2139a8dc-19bb-453f-a293-02b7f5944810\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.594806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l9x2\" (UniqueName: \"kubernetes.io/projected/584d122e-147a-48a3-993e-6ecd0ee9208a-kube-api-access-6l9x2\") pod \"584d122e-147a-48a3-993e-6ecd0ee9208a\" (UID: \"584d122e-147a-48a3-993e-6ecd0ee9208a\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.594838 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28t8n\" (UniqueName: \"kubernetes.io/projected/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-kube-api-access-28t8n\") pod \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.594951 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b9dc\" (UniqueName: \"kubernetes.io/projected/2139a8dc-19bb-453f-a293-02b7f5944810-kube-api-access-6b9dc\") pod \"2139a8dc-19bb-453f-a293-02b7f5944810\" (UID: \"2139a8dc-19bb-453f-a293-02b7f5944810\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.595072 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-operator-scripts\") pod \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\" (UID: \"13239d2e-d649-4a1c-aafc-fbdc135d7c2f\") " Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.596960 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "13239d2e-d649-4a1c-aafc-fbdc135d7c2f" (UID: "13239d2e-d649-4a1c-aafc-fbdc135d7c2f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.599242 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/584d122e-147a-48a3-993e-6ecd0ee9208a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "584d122e-147a-48a3-993e-6ecd0ee9208a" (UID: "584d122e-147a-48a3-993e-6ecd0ee9208a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.599571 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2139a8dc-19bb-453f-a293-02b7f5944810-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2139a8dc-19bb-453f-a293-02b7f5944810" (UID: "2139a8dc-19bb-453f-a293-02b7f5944810"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.601000 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x84z" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.604124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4x84z" event={"ID":"2139a8dc-19bb-453f-a293-02b7f5944810","Type":"ContainerDied","Data":"6d407af0a818bb2ff183a83537995f6224ba2aa73cff1b593dbd83b91e4016ed"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.604180 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d407af0a818bb2ff183a83537995f6224ba2aa73cff1b593dbd83b91e4016ed" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.607270 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584d122e-147a-48a3-993e-6ecd0ee9208a-kube-api-access-6l9x2" (OuterVolumeSpecName: "kube-api-access-6l9x2") pod "584d122e-147a-48a3-993e-6ecd0ee9208a" (UID: "584d122e-147a-48a3-993e-6ecd0ee9208a"). InnerVolumeSpecName "kube-api-access-6l9x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.607408 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-kube-api-access-28t8n" (OuterVolumeSpecName: "kube-api-access-28t8n") pod "13239d2e-d649-4a1c-aafc-fbdc135d7c2f" (UID: "13239d2e-d649-4a1c-aafc-fbdc135d7c2f"). InnerVolumeSpecName "kube-api-access-28t8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.611114 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.628485 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2139a8dc-19bb-453f-a293-02b7f5944810-kube-api-access-6b9dc" (OuterVolumeSpecName: "kube-api-access-6b9dc") pod "2139a8dc-19bb-453f-a293-02b7f5944810" (UID: "2139a8dc-19bb-453f-a293-02b7f5944810"). InnerVolumeSpecName "kube-api-access-6b9dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.639779 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"92c1ce5a-911d-4567-8d23-11777a6df443","Type":"ContainerStarted","Data":"03fe16447a853f4d3945d5618d9aa99ead7a5e4888f7d695b9dcd4332ab7d379"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.663014 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.692097 4715 scope.go:117] "RemoveContainer" containerID="34ebffad1ccd13f9dd6113bd6daea6476e89742be07724b139fd53e1c3592d0b" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693096 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693621 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584d122e-147a-48a3-993e-6ecd0ee9208a" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693640 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="584d122e-147a-48a3-993e-6ecd0ee9208a" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693667 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13239d2e-d649-4a1c-aafc-fbdc135d7c2f" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693676 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="13239d2e-d649-4a1c-aafc-fbdc135d7c2f" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693710 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-log" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693719 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-log" Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693735 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2139a8dc-19bb-453f-a293-02b7f5944810" containerName="mariadb-database-create" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693743 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2139a8dc-19bb-453f-a293-02b7f5944810" containerName="mariadb-database-create" Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693756 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49248bc-1dc1-45c2-b0a5-6886331d3c54" containerName="mariadb-database-create" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693764 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49248bc-1dc1-45c2-b0a5-6886331d3c54" containerName="mariadb-database-create" Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693782 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e7bb9b-d971-4641-b02d-0e17b0afe36f" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693790 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e7bb9b-d971-4641-b02d-0e17b0afe36f" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: E1204 14:20:12.693802 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-httpd" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.693810 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-httpd" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694672 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2139a8dc-19bb-453f-a293-02b7f5944810" containerName="mariadb-database-create" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694697 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="584d122e-147a-48a3-993e-6ecd0ee9208a" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694709 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e7bb9b-d971-4641-b02d-0e17b0afe36f" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694724 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="13239d2e-d649-4a1c-aafc-fbdc135d7c2f" containerName="mariadb-account-create-update" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694737 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-log" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694747 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49248bc-1dc1-45c2-b0a5-6886331d3c54" containerName="mariadb-database-create" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.694758 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" containerName="glance-httpd" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696168 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerStarted","Data":"bc37a5df3834b964b2ace0b3636efe5c36ed2f9512ea91bb6bf50a84f9b5f7f4"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696431 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696838 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696861 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d122e-147a-48a3-993e-6ecd0ee9208a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696870 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2139a8dc-19bb-453f-a293-02b7f5944810-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696878 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l9x2\" (UniqueName: \"kubernetes.io/projected/584d122e-147a-48a3-993e-6ecd0ee9208a-kube-api-access-6l9x2\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696892 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28t8n\" (UniqueName: \"kubernetes.io/projected/13239d2e-d649-4a1c-aafc-fbdc135d7c2f-kube-api-access-28t8n\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.696900 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b9dc\" (UniqueName: \"kubernetes.io/projected/2139a8dc-19bb-453f-a293-02b7f5944810-kube-api-access-6b9dc\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.701124 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.701354 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.704273 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-01e0-account-create-update-h84ff" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.704377 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-01e0-account-create-update-h84ff" event={"ID":"584d122e-147a-48a3-993e-6ecd0ee9208a","Type":"ContainerDied","Data":"2b02a262ff9daf6600450234fbb61c2d89e08f56ced9026a88cbcb431d97f1c9"} Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.704407 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b02a262ff9daf6600450234fbb61c2d89e08f56ced9026a88cbcb431d97f1c9" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.746284 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801109 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bea4af7-30c3-4278-99d7-01f1c604a89c-logs\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801530 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4tx\" (UniqueName: \"kubernetes.io/projected/3bea4af7-30c3-4278-99d7-01f1c604a89c-kube-api-access-dw4tx\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801552 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801610 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bea4af7-30c3-4278-99d7-01f1c604a89c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801754 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.801800 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.918330 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.918443 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.918502 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.918728 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bea4af7-30c3-4278-99d7-01f1c604a89c-logs\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.918934 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.919103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4tx\" (UniqueName: \"kubernetes.io/projected/3bea4af7-30c3-4278-99d7-01f1c604a89c-kube-api-access-dw4tx\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.919178 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.919320 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bea4af7-30c3-4278-99d7-01f1c604a89c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.921610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bea4af7-30c3-4278-99d7-01f1c604a89c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.922334 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bea4af7-30c3-4278-99d7-01f1c604a89c-logs\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.922792 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.948180 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.948683 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.952648 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.970585 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4tx\" (UniqueName: \"kubernetes.io/projected/3bea4af7-30c3-4278-99d7-01f1c604a89c-kube-api-access-dw4tx\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:12 crc kubenswrapper[4715]: I1204 14:20:12.976877 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bea4af7-30c3-4278-99d7-01f1c604a89c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.098412 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"3bea4af7-30c3-4278-99d7-01f1c604a89c\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.138593 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.217571 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d95418a-f1f5-4676-903e-67b2fee4417c" path="/var/lib/kubelet/pods/7d95418a-f1f5-4676-903e-67b2fee4417c/volumes" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.294147 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:13 crc kubenswrapper[4715]: E1204 14:20:13.318365 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2139a8dc_19bb_453f_a293_02b7f5944810.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13239d2e_d649_4a1c_aafc_fbdc135d7c2f.slice/crio-5bc002f7d3eea49a690646bdec63fcd1119f939c885700f7f7a2f0bc0c1ef5eb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13239d2e_d649_4a1c_aafc_fbdc135d7c2f.slice\": RecentStats: unable to find data in memory cache]" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.432601 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff544d83-9fa6-444c-bb1e-212318e8dc6a-operator-scripts\") pod \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.432707 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfvqf\" (UniqueName: \"kubernetes.io/projected/ff544d83-9fa6-444c-bb1e-212318e8dc6a-kube-api-access-bfvqf\") pod \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\" (UID: \"ff544d83-9fa6-444c-bb1e-212318e8dc6a\") " Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.435942 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff544d83-9fa6-444c-bb1e-212318e8dc6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff544d83-9fa6-444c-bb1e-212318e8dc6a" (UID: "ff544d83-9fa6-444c-bb1e-212318e8dc6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.439159 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff544d83-9fa6-444c-bb1e-212318e8dc6a-kube-api-access-bfvqf" (OuterVolumeSpecName: "kube-api-access-bfvqf") pod "ff544d83-9fa6-444c-bb1e-212318e8dc6a" (UID: "ff544d83-9fa6-444c-bb1e-212318e8dc6a"). InnerVolumeSpecName "kube-api-access-bfvqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.535086 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfvqf\" (UniqueName: \"kubernetes.io/projected/ff544d83-9fa6-444c-bb1e-212318e8dc6a-kube-api-access-bfvqf\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.535419 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff544d83-9fa6-444c-bb1e-212318e8dc6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.673250 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.720652 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"92c1ce5a-911d-4567-8d23-11777a6df443","Type":"ContainerStarted","Data":"0d07675496a7a4671dd6ac612a63087e6d65cfaf56d73b41b6a83fce78a5cdcd"} Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.727291 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbgxz" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.728049 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbgxz" event={"ID":"ff544d83-9fa6-444c-bb1e-212318e8dc6a","Type":"ContainerDied","Data":"09c546d6e71b7e440f67c0fda56877cb77057511626eee5633e6f11f359fbaea"} Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.728082 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09c546d6e71b7e440f67c0fda56877cb77057511626eee5633e6f11f359fbaea" Dec 04 14:20:13 crc kubenswrapper[4715]: I1204 14:20:13.730440 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3bea4af7-30c3-4278-99d7-01f1c604a89c","Type":"ContainerStarted","Data":"1b082d1058628b917a7bdf036fdbe722a8a76a6524f3432995332c8190d2f107"} Dec 04 14:20:14 crc kubenswrapper[4715]: I1204 14:20:14.561662 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:20:14 crc kubenswrapper[4715]: I1204 14:20:14.563308 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-64bbd97b75-ssj5n" Dec 04 14:20:14 crc kubenswrapper[4715]: I1204 14:20:14.751140 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"92c1ce5a-911d-4567-8d23-11777a6df443","Type":"ContainerStarted","Data":"6e9563ccbd7069cb9b2b8d89d2407dc6bf78493d4d2ee994e15434652f6c65d1"} Dec 04 14:20:14 crc kubenswrapper[4715]: I1204 14:20:14.756178 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3bea4af7-30c3-4278-99d7-01f1c604a89c","Type":"ContainerStarted","Data":"7ee86398cc2e1769b2e02f8364b532fcf6cd758a31b77ca11b44bb1e1b18d97d"} Dec 04 14:20:14 crc kubenswrapper[4715]: I1204 14:20:14.776658 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.776636573 podStartE2EDuration="4.776636573s" podCreationTimestamp="2025-12-04 14:20:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:20:14.772284675 +0000 UTC m=+1371.841002900" watchObservedRunningTime="2025-12-04 14:20:14.776636573 +0000 UTC m=+1371.845354788" Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.771285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerStarted","Data":"1a3df0c0ce7a8d39a2e6841d5400dee9b82cc30856ea3ba8b32e44cbfc4448df"} Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.771906 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.771385 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-central-agent" containerID="cri-o://1e630ce7abae93726166f066bace8ceed01d5e310f1daebdb2ce9780eaeebc7a" gracePeriod=30 Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.771662 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="sg-core" containerID="cri-o://bc37a5df3834b964b2ace0b3636efe5c36ed2f9512ea91bb6bf50a84f9b5f7f4" gracePeriod=30 Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.771645 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="proxy-httpd" containerID="cri-o://1a3df0c0ce7a8d39a2e6841d5400dee9b82cc30856ea3ba8b32e44cbfc4448df" gracePeriod=30 Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.771681 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-notification-agent" containerID="cri-o://b86ff9faafd6567d9e9147f726cf05b3e0ce4097006bd21594161baa66c853f8" gracePeriod=30 Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.775866 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3bea4af7-30c3-4278-99d7-01f1c604a89c","Type":"ContainerStarted","Data":"5f7501f7284cf6836ecfa43eedf763d00c033ac669d08563aa8772958cd1b577"} Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.815771 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.891239657 podStartE2EDuration="8.815743944s" podCreationTimestamp="2025-12-04 14:20:07 +0000 UTC" firstStartedPulling="2025-12-04 14:20:08.970835201 +0000 UTC m=+1366.039553416" lastFinishedPulling="2025-12-04 14:20:14.895339488 +0000 UTC m=+1371.964057703" observedRunningTime="2025-12-04 14:20:15.801627812 +0000 UTC m=+1372.870346047" watchObservedRunningTime="2025-12-04 14:20:15.815743944 +0000 UTC m=+1372.884462159" Dec 04 14:20:15 crc kubenswrapper[4715]: I1204 14:20:15.833278 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.833251228 podStartE2EDuration="3.833251228s" podCreationTimestamp="2025-12-04 14:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:20:15.828513 +0000 UTC m=+1372.897231215" watchObservedRunningTime="2025-12-04 14:20:15.833251228 +0000 UTC m=+1372.901969443" Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.798608 4715 generic.go:334] "Generic (PLEG): container finished" podID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerID="1a3df0c0ce7a8d39a2e6841d5400dee9b82cc30856ea3ba8b32e44cbfc4448df" exitCode=0 Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.798878 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerDied","Data":"1a3df0c0ce7a8d39a2e6841d5400dee9b82cc30856ea3ba8b32e44cbfc4448df"} Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.798955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerDied","Data":"bc37a5df3834b964b2ace0b3636efe5c36ed2f9512ea91bb6bf50a84f9b5f7f4"} Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.798900 4715 generic.go:334] "Generic (PLEG): container finished" podID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerID="bc37a5df3834b964b2ace0b3636efe5c36ed2f9512ea91bb6bf50a84f9b5f7f4" exitCode=2 Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.798999 4715 generic.go:334] "Generic (PLEG): container finished" podID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerID="b86ff9faafd6567d9e9147f726cf05b3e0ce4097006bd21594161baa66c853f8" exitCode=0 Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.799019 4715 generic.go:334] "Generic (PLEG): container finished" podID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerID="1e630ce7abae93726166f066bace8ceed01d5e310f1daebdb2ce9780eaeebc7a" exitCode=0 Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.799092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerDied","Data":"b86ff9faafd6567d9e9147f726cf05b3e0ce4097006bd21594161baa66c853f8"} Dec 04 14:20:16 crc kubenswrapper[4715]: I1204 14:20:16.799133 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerDied","Data":"1e630ce7abae93726166f066bace8ceed01d5e310f1daebdb2ce9780eaeebc7a"} Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.149936 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.210745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-run-httpd\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.210800 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-config-data\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.210903 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-sg-core-conf-yaml\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.210959 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-combined-ca-bundle\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.211112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-scripts\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.211147 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-log-httpd\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.211290 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgmdh\" (UniqueName: \"kubernetes.io/projected/b80b13f4-50d2-4586-a7c5-5cf18273bb48-kube-api-access-tgmdh\") pod \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\" (UID: \"b80b13f4-50d2-4586-a7c5-5cf18273bb48\") " Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.214215 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.214381 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.223129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80b13f4-50d2-4586-a7c5-5cf18273bb48-kube-api-access-tgmdh" (OuterVolumeSpecName: "kube-api-access-tgmdh") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "kube-api-access-tgmdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.239613 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-scripts" (OuterVolumeSpecName: "scripts") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.261504 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.317814 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.317856 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.317871 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.317883 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgmdh\" (UniqueName: \"kubernetes.io/projected/b80b13f4-50d2-4586-a7c5-5cf18273bb48-kube-api-access-tgmdh\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.317902 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b80b13f4-50d2-4586-a7c5-5cf18273bb48-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.326250 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.329831 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-config-data" (OuterVolumeSpecName: "config-data") pod "b80b13f4-50d2-4586-a7c5-5cf18273bb48" (UID: "b80b13f4-50d2-4586-a7c5-5cf18273bb48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.419103 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.419136 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b13f4-50d2-4586-a7c5-5cf18273bb48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.765540 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fkv2x"] Dec 04 14:20:17 crc kubenswrapper[4715]: E1204 14:20:17.765981 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-notification-agent" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766002 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-notification-agent" Dec 04 14:20:17 crc kubenswrapper[4715]: E1204 14:20:17.766029 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff544d83-9fa6-444c-bb1e-212318e8dc6a" containerName="mariadb-database-create" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766059 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff544d83-9fa6-444c-bb1e-212318e8dc6a" containerName="mariadb-database-create" Dec 04 14:20:17 crc kubenswrapper[4715]: E1204 14:20:17.766075 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="sg-core" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766084 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="sg-core" Dec 04 14:20:17 crc kubenswrapper[4715]: E1204 14:20:17.766104 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="proxy-httpd" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766112 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="proxy-httpd" Dec 04 14:20:17 crc kubenswrapper[4715]: E1204 14:20:17.766128 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-central-agent" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766136 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-central-agent" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766334 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-notification-agent" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766349 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff544d83-9fa6-444c-bb1e-212318e8dc6a" containerName="mariadb-database-create" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766363 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="ceilometer-central-agent" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766372 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="sg-core" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.766382 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" containerName="proxy-httpd" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.767065 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.770978 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-dmzb4" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.771185 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.771337 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.795630 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fkv2x"] Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.831482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzwwk\" (UniqueName: \"kubernetes.io/projected/d036e63f-d6a8-47cc-a19f-71a1dccf7698-kube-api-access-fzwwk\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.831557 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-scripts\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.831597 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-config-data\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.831625 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.834843 4715 generic.go:334] "Generic (PLEG): container finished" podID="88520da5-8fee-4a16-9396-2280c468c41a" containerID="1341f08949be200925226253e4ab03bfe80d9e1c59dcf0907a757e3dab2e4ec3" exitCode=137 Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.834945 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754dbdc69b-nwpzv" event={"ID":"88520da5-8fee-4a16-9396-2280c468c41a","Type":"ContainerDied","Data":"1341f08949be200925226253e4ab03bfe80d9e1c59dcf0907a757e3dab2e4ec3"} Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.905301 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b80b13f4-50d2-4586-a7c5-5cf18273bb48","Type":"ContainerDied","Data":"0d9f99f4811808b653a6e6241de5333a069afe6853e3decf84f74e97c01bbff8"} Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.905355 4715 scope.go:117] "RemoveContainer" containerID="1a3df0c0ce7a8d39a2e6841d5400dee9b82cc30856ea3ba8b32e44cbfc4448df" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.905497 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.934005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-config-data\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.934103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.934384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwwk\" (UniqueName: \"kubernetes.io/projected/d036e63f-d6a8-47cc-a19f-71a1dccf7698-kube-api-access-fzwwk\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.934463 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-scripts\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.941375 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-config-data\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.943879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.945572 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-scripts\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.949385 4715 scope.go:117] "RemoveContainer" containerID="bc37a5df3834b964b2ace0b3636efe5c36ed2f9512ea91bb6bf50a84f9b5f7f4" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.952809 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.953711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwwk\" (UniqueName: \"kubernetes.io/projected/d036e63f-d6a8-47cc-a19f-71a1dccf7698-kube-api-access-fzwwk\") pod \"nova-cell0-conductor-db-sync-fkv2x\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.963370 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.994726 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.995118 4715 scope.go:117] "RemoveContainer" containerID="b86ff9faafd6567d9e9147f726cf05b3e0ce4097006bd21594161baa66c853f8" Dec 04 14:20:17 crc kubenswrapper[4715]: I1204 14:20:17.998342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.002472 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.002782 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.007119 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-run-httpd\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036403 4715 scope.go:117] "RemoveContainer" containerID="1e630ce7abae93726166f066bace8ceed01d5e310f1daebdb2ce9780eaeebc7a" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036513 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-scripts\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036555 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-log-httpd\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036575 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036595 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036630 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-config-data\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.036653 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8qkk\" (UniqueName: \"kubernetes.io/projected/2d9d8267-65e3-49b6-99fd-70082e20ba47-kube-api-access-c8qkk\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.123480 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.138832 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-log-httpd\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.138879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.138908 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.138963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-config-data\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.138992 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8qkk\" (UniqueName: \"kubernetes.io/projected/2d9d8267-65e3-49b6-99fd-70082e20ba47-kube-api-access-c8qkk\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.139071 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-run-httpd\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.139250 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-scripts\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.139449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-log-httpd\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.141951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-run-httpd\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.144910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.145479 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-config-data\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.146845 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-scripts\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.156680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.160681 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8qkk\" (UniqueName: \"kubernetes.io/projected/2d9d8267-65e3-49b6-99fd-70082e20ba47-kube-api-access-c8qkk\") pod \"ceilometer-0\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.327966 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.617685 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fkv2x"] Dec 04 14:20:18 crc kubenswrapper[4715]: W1204 14:20:18.624432 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd036e63f_d6a8_47cc_a19f_71a1dccf7698.slice/crio-645881d23dcf30302c8c775c02ccca0aec7f30a4eeb1860ebd2ba9a0affbd2ca WatchSource:0}: Error finding container 645881d23dcf30302c8c775c02ccca0aec7f30a4eeb1860ebd2ba9a0affbd2ca: Status 404 returned error can't find the container with id 645881d23dcf30302c8c775c02ccca0aec7f30a4eeb1860ebd2ba9a0affbd2ca Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.770578 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754dbdc69b-nwpzv" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.802199 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.916749 4715 generic.go:334] "Generic (PLEG): container finished" podID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerID="dd7b9e026fbce19f88bd2063857eba783cb3f3f66bde675059e0b4d466564778" exitCode=0 Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.916832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7bd77bfb-nlk9m" event={"ID":"d49eba49-c182-4ec6-9a61-d14cddf541fd","Type":"ContainerDied","Data":"dd7b9e026fbce19f88bd2063857eba783cb3f3f66bde675059e0b4d466564778"} Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.920293 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" event={"ID":"d036e63f-d6a8-47cc-a19f-71a1dccf7698","Type":"ContainerStarted","Data":"645881d23dcf30302c8c775c02ccca0aec7f30a4eeb1860ebd2ba9a0affbd2ca"} Dec 04 14:20:18 crc kubenswrapper[4715]: I1204 14:20:18.921530 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerStarted","Data":"6dfa4b99ca7bb4cf9ec3fc3b8934c8abe0742a0ced841a27593f9f03cdd25d2c"} Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.216769 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80b13f4-50d2-4586-a7c5-5cf18273bb48" path="/var/lib/kubelet/pods/b80b13f4-50d2-4586-a7c5-5cf18273bb48/volumes" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.304684 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.362723 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88520da5-8fee-4a16-9396-2280c468c41a-logs\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.362834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-secret-key\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.362869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-scripts\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.362894 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-tls-certs\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.363394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88520da5-8fee-4a16-9396-2280c468c41a-logs" (OuterVolumeSpecName: "logs") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.363454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpp92\" (UniqueName: \"kubernetes.io/projected/88520da5-8fee-4a16-9396-2280c468c41a-kube-api-access-xpp92\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.363506 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-config-data\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.363608 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-combined-ca-bundle\") pod \"88520da5-8fee-4a16-9396-2280c468c41a\" (UID: \"88520da5-8fee-4a16-9396-2280c468c41a\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.364170 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88520da5-8fee-4a16-9396-2280c468c41a-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.369222 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88520da5-8fee-4a16-9396-2280c468c41a-kube-api-access-xpp92" (OuterVolumeSpecName: "kube-api-access-xpp92") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "kube-api-access-xpp92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.378913 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.422977 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.432227 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-config-data" (OuterVolumeSpecName: "config-data") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.432508 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-scripts" (OuterVolumeSpecName: "scripts") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.432511 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.448789 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "88520da5-8fee-4a16-9396-2280c468c41a" (UID: "88520da5-8fee-4a16-9396-2280c468c41a"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.467214 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-config\") pod \"d49eba49-c182-4ec6-9a61-d14cddf541fd\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.467292 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-combined-ca-bundle\") pod \"d49eba49-c182-4ec6-9a61-d14cddf541fd\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.467331 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-ovndb-tls-certs\") pod \"d49eba49-c182-4ec6-9a61-d14cddf541fd\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.467462 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-httpd-config\") pod \"d49eba49-c182-4ec6-9a61-d14cddf541fd\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.467597 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlh4v\" (UniqueName: \"kubernetes.io/projected/d49eba49-c182-4ec6-9a61-d14cddf541fd-kube-api-access-dlh4v\") pod \"d49eba49-c182-4ec6-9a61-d14cddf541fd\" (UID: \"d49eba49-c182-4ec6-9a61-d14cddf541fd\") " Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.468222 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpp92\" (UniqueName: \"kubernetes.io/projected/88520da5-8fee-4a16-9396-2280c468c41a-kube-api-access-xpp92\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.468247 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.468263 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.468275 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.468286 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88520da5-8fee-4a16-9396-2280c468c41a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.468297 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/88520da5-8fee-4a16-9396-2280c468c41a-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.471968 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d49eba49-c182-4ec6-9a61-d14cddf541fd" (UID: "d49eba49-c182-4ec6-9a61-d14cddf541fd"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.489378 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49eba49-c182-4ec6-9a61-d14cddf541fd-kube-api-access-dlh4v" (OuterVolumeSpecName: "kube-api-access-dlh4v") pod "d49eba49-c182-4ec6-9a61-d14cddf541fd" (UID: "d49eba49-c182-4ec6-9a61-d14cddf541fd"). InnerVolumeSpecName "kube-api-access-dlh4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.570492 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.570527 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlh4v\" (UniqueName: \"kubernetes.io/projected/d49eba49-c182-4ec6-9a61-d14cddf541fd-kube-api-access-dlh4v\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.581208 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d49eba49-c182-4ec6-9a61-d14cddf541fd" (UID: "d49eba49-c182-4ec6-9a61-d14cddf541fd"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.589871 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-config" (OuterVolumeSpecName: "config") pod "d49eba49-c182-4ec6-9a61-d14cddf541fd" (UID: "d49eba49-c182-4ec6-9a61-d14cddf541fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.600848 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d49eba49-c182-4ec6-9a61-d14cddf541fd" (UID: "d49eba49-c182-4ec6-9a61-d14cddf541fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.672993 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.673112 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.673128 4715 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49eba49-c182-4ec6-9a61-d14cddf541fd-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.934504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerStarted","Data":"ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1"} Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.937894 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7bd77bfb-nlk9m" event={"ID":"d49eba49-c182-4ec6-9a61-d14cddf541fd","Type":"ContainerDied","Data":"32d6029c149005d6daaf527ea08fe637a30db48ac7d8123e46d08170e7274461"} Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.937936 4715 scope.go:117] "RemoveContainer" containerID="0f42ca9a061f45d393889a4cfa54ca235d7648375ad5d892f8dc431de02d3d95" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.937948 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d7bd77bfb-nlk9m" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.940326 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754dbdc69b-nwpzv" event={"ID":"88520da5-8fee-4a16-9396-2280c468c41a","Type":"ContainerDied","Data":"9b8b7a0b24c121328511b9162416397e6e369f78ee03d2c74f62269c993848c2"} Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.940409 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754dbdc69b-nwpzv" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.967257 4715 scope.go:117] "RemoveContainer" containerID="dd7b9e026fbce19f88bd2063857eba783cb3f3f66bde675059e0b4d466564778" Dec 04 14:20:19 crc kubenswrapper[4715]: I1204 14:20:19.997584 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-754dbdc69b-nwpzv"] Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.010720 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-754dbdc69b-nwpzv"] Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.012618 4715 scope.go:117] "RemoveContainer" containerID="f1297a46867ceb825461ae2c754dc4d5b3ba962c40fa5376fd6f94ccf7879eb0" Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.025509 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d7bd77bfb-nlk9m"] Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.036784 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d7bd77bfb-nlk9m"] Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.207060 4715 scope.go:117] "RemoveContainer" containerID="1341f08949be200925226253e4ab03bfe80d9e1c59dcf0907a757e3dab2e4ec3" Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.952672 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72","Type":"ContainerStarted","Data":"07e5ec1c23af9fe9cfa29ff1523189415a278f3ab9f3a7d162fb8da3b83c3131"} Dec 04 14:20:20 crc kubenswrapper[4715]: I1204 14:20:20.980166 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.358156992 podStartE2EDuration="32.980135204s" podCreationTimestamp="2025-12-04 14:19:48 +0000 UTC" firstStartedPulling="2025-12-04 14:19:49.744952225 +0000 UTC m=+1346.813670440" lastFinishedPulling="2025-12-04 14:20:20.366930437 +0000 UTC m=+1377.435648652" observedRunningTime="2025-12-04 14:20:20.97558295 +0000 UTC m=+1378.044301165" watchObservedRunningTime="2025-12-04 14:20:20.980135204 +0000 UTC m=+1378.048853419" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.102158 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.102458 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.140076 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.155575 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.197766 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88520da5-8fee-4a16-9396-2280c468c41a" path="/var/lib/kubelet/pods/88520da5-8fee-4a16-9396-2280c468c41a/volumes" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.198715 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" path="/var/lib/kubelet/pods/d49eba49-c182-4ec6-9a61-d14cddf541fd/volumes" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.983947 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerStarted","Data":"9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06"} Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.984432 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:20:21 crc kubenswrapper[4715]: I1204 14:20:21.984478 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:20:23 crc kubenswrapper[4715]: I1204 14:20:23.139084 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:23 crc kubenswrapper[4715]: I1204 14:20:23.139417 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:23 crc kubenswrapper[4715]: I1204 14:20:23.264178 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:23 crc kubenswrapper[4715]: I1204 14:20:23.264963 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:24 crc kubenswrapper[4715]: I1204 14:20:24.002803 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:24 crc kubenswrapper[4715]: I1204 14:20:24.004305 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:24 crc kubenswrapper[4715]: I1204 14:20:24.218122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:20:24 crc kubenswrapper[4715]: I1204 14:20:24.218262 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 14:20:24 crc kubenswrapper[4715]: I1204 14:20:24.221780 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:20:26 crc kubenswrapper[4715]: I1204 14:20:26.137409 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:26 crc kubenswrapper[4715]: I1204 14:20:26.138101 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 14:20:26 crc kubenswrapper[4715]: I1204 14:20:26.138546 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:20:27 crc kubenswrapper[4715]: I1204 14:20:27.254424 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:29 crc kubenswrapper[4715]: I1204 14:20:29.081849 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerStarted","Data":"daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87"} Dec 04 14:20:30 crc kubenswrapper[4715]: I1204 14:20:30.095395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" event={"ID":"d036e63f-d6a8-47cc-a19f-71a1dccf7698","Type":"ContainerStarted","Data":"d6657b9b09615afca6ca840be9dbe4c674110c0a7d955559eed125246a562fd8"} Dec 04 14:20:30 crc kubenswrapper[4715]: I1204 14:20:30.111424 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" podStartSLOduration=2.660360393 podStartE2EDuration="13.111206648s" podCreationTimestamp="2025-12-04 14:20:17 +0000 UTC" firstStartedPulling="2025-12-04 14:20:18.628381005 +0000 UTC m=+1375.697099220" lastFinishedPulling="2025-12-04 14:20:29.07922727 +0000 UTC m=+1386.147945475" observedRunningTime="2025-12-04 14:20:30.109483321 +0000 UTC m=+1387.178201536" watchObservedRunningTime="2025-12-04 14:20:30.111206648 +0000 UTC m=+1387.179924863" Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.107834 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerStarted","Data":"8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352"} Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.108021 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-central-agent" containerID="cri-o://ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1" gracePeriod=30 Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.108109 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="proxy-httpd" containerID="cri-o://8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352" gracePeriod=30 Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.108263 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.108132 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="sg-core" containerID="cri-o://daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87" gracePeriod=30 Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.108136 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-notification-agent" containerID="cri-o://9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06" gracePeriod=30 Dec 04 14:20:31 crc kubenswrapper[4715]: I1204 14:20:31.129976 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.71013939 podStartE2EDuration="14.129956657s" podCreationTimestamp="2025-12-04 14:20:17 +0000 UTC" firstStartedPulling="2025-12-04 14:20:18.808540524 +0000 UTC m=+1375.877258739" lastFinishedPulling="2025-12-04 14:20:30.228357791 +0000 UTC m=+1387.297076006" observedRunningTime="2025-12-04 14:20:31.126697779 +0000 UTC m=+1388.195415994" watchObservedRunningTime="2025-12-04 14:20:31.129956657 +0000 UTC m=+1388.198674872" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.119939 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerID="8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352" exitCode=0 Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.120203 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerID="daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87" exitCode=2 Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.120212 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerID="ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1" exitCode=0 Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.120020 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerDied","Data":"8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352"} Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.120248 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerDied","Data":"daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87"} Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.120262 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerDied","Data":"ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1"} Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.704181 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.770900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-config-data\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.771257 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-sg-core-conf-yaml\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.771539 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8qkk\" (UniqueName: \"kubernetes.io/projected/2d9d8267-65e3-49b6-99fd-70082e20ba47-kube-api-access-c8qkk\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.771660 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-run-httpd\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.771793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-log-httpd\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.771920 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-combined-ca-bundle\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.772129 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-scripts\") pod \"2d9d8267-65e3-49b6-99fd-70082e20ba47\" (UID: \"2d9d8267-65e3-49b6-99fd-70082e20ba47\") " Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.771980 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.772197 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.772874 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.773702 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9d8267-65e3-49b6-99fd-70082e20ba47-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.790021 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-scripts" (OuterVolumeSpecName: "scripts") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.790268 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9d8267-65e3-49b6-99fd-70082e20ba47-kube-api-access-c8qkk" (OuterVolumeSpecName: "kube-api-access-c8qkk") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "kube-api-access-c8qkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.808181 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.851998 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.875277 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.875325 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.875339 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8qkk\" (UniqueName: \"kubernetes.io/projected/2d9d8267-65e3-49b6-99fd-70082e20ba47-kube-api-access-c8qkk\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.875352 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.889693 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-config-data" (OuterVolumeSpecName: "config-data") pod "2d9d8267-65e3-49b6-99fd-70082e20ba47" (UID: "2d9d8267-65e3-49b6-99fd-70082e20ba47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:32 crc kubenswrapper[4715]: I1204 14:20:32.977794 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9d8267-65e3-49b6-99fd-70082e20ba47-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.134443 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerID="9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06" exitCode=0 Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.134498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerDied","Data":"9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06"} Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.134532 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9d8267-65e3-49b6-99fd-70082e20ba47","Type":"ContainerDied","Data":"6dfa4b99ca7bb4cf9ec3fc3b8934c8abe0742a0ced841a27593f9f03cdd25d2c"} Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.134557 4715 scope.go:117] "RemoveContainer" containerID="8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.134720 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.174317 4715 scope.go:117] "RemoveContainer" containerID="daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.175290 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210129 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210185 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210560 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-api" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210582 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-api" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210603 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-central-agent" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210612 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-central-agent" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210623 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210631 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210648 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="sg-core" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210656 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="sg-core" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210668 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon-log" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210675 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon-log" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210686 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-notification-agent" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210693 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-notification-agent" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210706 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-httpd" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210713 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-httpd" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.210737 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="proxy-httpd" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210746 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="proxy-httpd" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.210958 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-notification-agent" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211023 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="ceilometer-central-agent" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211056 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-httpd" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211067 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon-log" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211082 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="sg-core" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211092 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" containerName="proxy-httpd" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211108 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="88520da5-8fee-4a16-9396-2280c468c41a" containerName="horizon" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.211123 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49eba49-c182-4ec6-9a61-d14cddf541fd" containerName="neutron-api" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.213144 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.221030 4715 scope.go:117] "RemoveContainer" containerID="9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.221396 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.221549 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.229784 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.247016 4715 scope.go:117] "RemoveContainer" containerID="ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.282218 4715 scope.go:117] "RemoveContainer" containerID="8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.282757 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352\": container with ID starting with 8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352 not found: ID does not exist" containerID="8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.282880 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352"} err="failed to get container status \"8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352\": rpc error: code = NotFound desc = could not find container \"8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352\": container with ID starting with 8f5a9ddb4d1980fde136514e84203a68ccc053845e62e2f5fd98984746924352 not found: ID does not exist" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.283024 4715 scope.go:117] "RemoveContainer" containerID="daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.283472 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87\": container with ID starting with daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87 not found: ID does not exist" containerID="daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.283493 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87"} err="failed to get container status \"daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87\": rpc error: code = NotFound desc = could not find container \"daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87\": container with ID starting with daebfd5e214756d0a17e208ff79bf2093eaf809e456cc4b6f09d5b7ea7c26f87 not found: ID does not exist" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.283506 4715 scope.go:117] "RemoveContainer" containerID="9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.283728 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06\": container with ID starting with 9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06 not found: ID does not exist" containerID="9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.283838 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06"} err="failed to get container status \"9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06\": rpc error: code = NotFound desc = could not find container \"9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06\": container with ID starting with 9854271b9df6c8e9ef252da2a597ac27b507ac8229c2a4b241d6c44e68ec4a06 not found: ID does not exist" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.283911 4715 scope.go:117] "RemoveContainer" containerID="ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1" Dec 04 14:20:33 crc kubenswrapper[4715]: E1204 14:20:33.284185 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1\": container with ID starting with ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1 not found: ID does not exist" containerID="ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284205 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1"} err="failed to get container status \"ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1\": rpc error: code = NotFound desc = could not find container \"ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1\": container with ID starting with ee524f411d4675cee9162999d01805b9cd3b98b6c6ddbbbb0285b16daedd46a1 not found: ID does not exist" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284806 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrzf\" (UniqueName: \"kubernetes.io/projected/a3fd4198-2135-4d50-8857-21055b3f99b0-kube-api-access-ctrzf\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284847 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284891 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-run-httpd\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284919 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-scripts\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.284978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-config-data\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.285218 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-log-httpd\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.386609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-log-httpd\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387113 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctrzf\" (UniqueName: \"kubernetes.io/projected/a3fd4198-2135-4d50-8857-21055b3f99b0-kube-api-access-ctrzf\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387257 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-run-httpd\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-scripts\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387651 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-run-httpd\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387758 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-config-data\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.387359 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-log-httpd\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.392356 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.392434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-config-data\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.392618 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.392916 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-scripts\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.406781 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctrzf\" (UniqueName: \"kubernetes.io/projected/a3fd4198-2135-4d50-8857-21055b3f99b0-kube-api-access-ctrzf\") pod \"ceilometer-0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " pod="openstack/ceilometer-0" Dec 04 14:20:33 crc kubenswrapper[4715]: I1204 14:20:33.556885 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:34 crc kubenswrapper[4715]: I1204 14:20:34.024028 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:34 crc kubenswrapper[4715]: I1204 14:20:34.145350 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerStarted","Data":"c96feeb09b85ec0891a416caaf26f2ea10b271a366583603ee7b00cdd100514d"} Dec 04 14:20:35 crc kubenswrapper[4715]: I1204 14:20:35.193241 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d9d8267-65e3-49b6-99fd-70082e20ba47" path="/var/lib/kubelet/pods/2d9d8267-65e3-49b6-99fd-70082e20ba47/volumes" Dec 04 14:20:37 crc kubenswrapper[4715]: I1204 14:20:37.174022 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerStarted","Data":"a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe"} Dec 04 14:20:38 crc kubenswrapper[4715]: I1204 14:20:38.188054 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerStarted","Data":"1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1"} Dec 04 14:20:39 crc kubenswrapper[4715]: I1204 14:20:39.199129 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerStarted","Data":"a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0"} Dec 04 14:20:40 crc kubenswrapper[4715]: I1204 14:20:40.218241 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerStarted","Data":"5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac"} Dec 04 14:20:40 crc kubenswrapper[4715]: I1204 14:20:40.218561 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:20:40 crc kubenswrapper[4715]: I1204 14:20:40.238781 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.663983613 podStartE2EDuration="7.238753088s" podCreationTimestamp="2025-12-04 14:20:33 +0000 UTC" firstStartedPulling="2025-12-04 14:20:34.043068278 +0000 UTC m=+1391.111786493" lastFinishedPulling="2025-12-04 14:20:39.617837753 +0000 UTC m=+1396.686555968" observedRunningTime="2025-12-04 14:20:40.237909605 +0000 UTC m=+1397.306627840" watchObservedRunningTime="2025-12-04 14:20:40.238753088 +0000 UTC m=+1397.307471313" Dec 04 14:20:43 crc kubenswrapper[4715]: I1204 14:20:43.737217 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:43 crc kubenswrapper[4715]: I1204 14:20:43.738095 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-central-agent" containerID="cri-o://a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe" gracePeriod=30 Dec 04 14:20:43 crc kubenswrapper[4715]: I1204 14:20:43.738183 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-notification-agent" containerID="cri-o://1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1" gracePeriod=30 Dec 04 14:20:43 crc kubenswrapper[4715]: I1204 14:20:43.738183 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="proxy-httpd" containerID="cri-o://5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac" gracePeriod=30 Dec 04 14:20:43 crc kubenswrapper[4715]: I1204 14:20:43.738166 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="sg-core" containerID="cri-o://a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0" gracePeriod=30 Dec 04 14:20:44 crc kubenswrapper[4715]: I1204 14:20:44.275426 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerID="5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac" exitCode=0 Dec 04 14:20:44 crc kubenswrapper[4715]: I1204 14:20:44.275740 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerID="a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0" exitCode=2 Dec 04 14:20:44 crc kubenswrapper[4715]: I1204 14:20:44.275753 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerID="1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1" exitCode=0 Dec 04 14:20:44 crc kubenswrapper[4715]: I1204 14:20:44.275779 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerDied","Data":"5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac"} Dec 04 14:20:44 crc kubenswrapper[4715]: I1204 14:20:44.275811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerDied","Data":"a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0"} Dec 04 14:20:44 crc kubenswrapper[4715]: I1204 14:20:44.275826 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerDied","Data":"1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1"} Dec 04 14:20:45 crc kubenswrapper[4715]: I1204 14:20:45.300545 4715 generic.go:334] "Generic (PLEG): container finished" podID="d036e63f-d6a8-47cc-a19f-71a1dccf7698" containerID="d6657b9b09615afca6ca840be9dbe4c674110c0a7d955559eed125246a562fd8" exitCode=0 Dec 04 14:20:45 crc kubenswrapper[4715]: I1204 14:20:45.300917 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" event={"ID":"d036e63f-d6a8-47cc-a19f-71a1dccf7698","Type":"ContainerDied","Data":"d6657b9b09615afca6ca840be9dbe4c674110c0a7d955559eed125246a562fd8"} Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.689244 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.789828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-combined-ca-bundle\") pod \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.789951 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzwwk\" (UniqueName: \"kubernetes.io/projected/d036e63f-d6a8-47cc-a19f-71a1dccf7698-kube-api-access-fzwwk\") pod \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.790183 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-config-data\") pod \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.790329 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-scripts\") pod \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\" (UID: \"d036e63f-d6a8-47cc-a19f-71a1dccf7698\") " Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.795744 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d036e63f-d6a8-47cc-a19f-71a1dccf7698-kube-api-access-fzwwk" (OuterVolumeSpecName: "kube-api-access-fzwwk") pod "d036e63f-d6a8-47cc-a19f-71a1dccf7698" (UID: "d036e63f-d6a8-47cc-a19f-71a1dccf7698"). InnerVolumeSpecName "kube-api-access-fzwwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.796101 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-scripts" (OuterVolumeSpecName: "scripts") pod "d036e63f-d6a8-47cc-a19f-71a1dccf7698" (UID: "d036e63f-d6a8-47cc-a19f-71a1dccf7698"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.816171 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-config-data" (OuterVolumeSpecName: "config-data") pod "d036e63f-d6a8-47cc-a19f-71a1dccf7698" (UID: "d036e63f-d6a8-47cc-a19f-71a1dccf7698"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.824848 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d036e63f-d6a8-47cc-a19f-71a1dccf7698" (UID: "d036e63f-d6a8-47cc-a19f-71a1dccf7698"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.893158 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.893201 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.893215 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzwwk\" (UniqueName: \"kubernetes.io/projected/d036e63f-d6a8-47cc-a19f-71a1dccf7698-kube-api-access-fzwwk\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:46 crc kubenswrapper[4715]: I1204 14:20:46.893225 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d036e63f-d6a8-47cc-a19f-71a1dccf7698-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.320236 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" event={"ID":"d036e63f-d6a8-47cc-a19f-71a1dccf7698","Type":"ContainerDied","Data":"645881d23dcf30302c8c775c02ccca0aec7f30a4eeb1860ebd2ba9a0affbd2ca"} Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.320294 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fkv2x" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.320303 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="645881d23dcf30302c8c775c02ccca0aec7f30a4eeb1860ebd2ba9a0affbd2ca" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.430439 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:20:47 crc kubenswrapper[4715]: E1204 14:20:47.430888 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d036e63f-d6a8-47cc-a19f-71a1dccf7698" containerName="nova-cell0-conductor-db-sync" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.430917 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d036e63f-d6a8-47cc-a19f-71a1dccf7698" containerName="nova-cell0-conductor-db-sync" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.431133 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d036e63f-d6a8-47cc-a19f-71a1dccf7698" containerName="nova-cell0-conductor-db-sync" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.431765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.436143 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-dmzb4" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.436694 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.444746 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.503612 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.503682 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.503709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pr46\" (UniqueName: \"kubernetes.io/projected/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-kube-api-access-4pr46\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.605438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.605523 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.605553 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pr46\" (UniqueName: \"kubernetes.io/projected/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-kube-api-access-4pr46\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.610438 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.610720 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.623495 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pr46\" (UniqueName: \"kubernetes.io/projected/02266663-d1a7-4968-ab1e-e4bb6b4a9b21-kube-api-access-4pr46\") pod \"nova-cell0-conductor-0\" (UID: \"02266663-d1a7-4968-ab1e-e4bb6b4a9b21\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:47 crc kubenswrapper[4715]: I1204 14:20:47.752052 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.193707 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:20:48 crc kubenswrapper[4715]: W1204 14:20:48.208971 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02266663_d1a7_4968_ab1e_e4bb6b4a9b21.slice/crio-7b81039a21655f99ef84e89fd7a72a05a0eef2e077e1f7f40222daca4b67be01 WatchSource:0}: Error finding container 7b81039a21655f99ef84e89fd7a72a05a0eef2e077e1f7f40222daca4b67be01: Status 404 returned error can't find the container with id 7b81039a21655f99ef84e89fd7a72a05a0eef2e077e1f7f40222daca4b67be01 Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.316552 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.329077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"02266663-d1a7-4968-ab1e-e4bb6b4a9b21","Type":"ContainerStarted","Data":"7b81039a21655f99ef84e89fd7a72a05a0eef2e077e1f7f40222daca4b67be01"} Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.335209 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerID="a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe" exitCode=0 Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.335258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerDied","Data":"a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe"} Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.335289 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3fd4198-2135-4d50-8857-21055b3f99b0","Type":"ContainerDied","Data":"c96feeb09b85ec0891a416caaf26f2ea10b271a366583603ee7b00cdd100514d"} Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.335309 4715 scope.go:117] "RemoveContainer" containerID="5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.335469 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.373771 4715 scope.go:117] "RemoveContainer" containerID="a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.395195 4715 scope.go:117] "RemoveContainer" containerID="1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.420446 4715 scope.go:117] "RemoveContainer" containerID="a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.421301 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-sg-core-conf-yaml\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.421355 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-log-httpd\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.422183 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.422302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-config-data\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.422717 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-combined-ca-bundle\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.422788 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-scripts\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.422832 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-run-httpd\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.422864 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctrzf\" (UniqueName: \"kubernetes.io/projected/a3fd4198-2135-4d50-8857-21055b3f99b0-kube-api-access-ctrzf\") pod \"a3fd4198-2135-4d50-8857-21055b3f99b0\" (UID: \"a3fd4198-2135-4d50-8857-21055b3f99b0\") " Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.423471 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.424516 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.424540 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4198-2135-4d50-8857-21055b3f99b0-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.427052 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3fd4198-2135-4d50-8857-21055b3f99b0-kube-api-access-ctrzf" (OuterVolumeSpecName: "kube-api-access-ctrzf") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "kube-api-access-ctrzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.427087 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-scripts" (OuterVolumeSpecName: "scripts") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.451317 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.459474 4715 scope.go:117] "RemoveContainer" containerID="5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.459962 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac\": container with ID starting with 5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac not found: ID does not exist" containerID="5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.460025 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac"} err="failed to get container status \"5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac\": rpc error: code = NotFound desc = could not find container \"5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac\": container with ID starting with 5e26573f70665f8f8d0add63230835ce3f5c3dd213b58bc1a50782c76d53c7ac not found: ID does not exist" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.460276 4715 scope.go:117] "RemoveContainer" containerID="a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.460934 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0\": container with ID starting with a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0 not found: ID does not exist" containerID="a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.460973 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0"} err="failed to get container status \"a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0\": rpc error: code = NotFound desc = could not find container \"a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0\": container with ID starting with a9d8f6188988d7dfb6847dd7c92e156f5cb203f5f863fc8afea379bc23b65cc0 not found: ID does not exist" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.460997 4715 scope.go:117] "RemoveContainer" containerID="1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.461603 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1\": container with ID starting with 1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1 not found: ID does not exist" containerID="1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.461679 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1"} err="failed to get container status \"1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1\": rpc error: code = NotFound desc = could not find container \"1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1\": container with ID starting with 1608641138a12db66c315b82420243ef3f936df903921f8a9b9f65eea54900e1 not found: ID does not exist" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.461716 4715 scope.go:117] "RemoveContainer" containerID="a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.462311 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe\": container with ID starting with a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe not found: ID does not exist" containerID="a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.462343 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe"} err="failed to get container status \"a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe\": rpc error: code = NotFound desc = could not find container \"a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe\": container with ID starting with a5ef4a1d1cb5f6d62a8fbcffc7cd1c0e1f8b4ffd8b50837f017519b5ecf518fe not found: ID does not exist" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.516435 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.526721 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.526758 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.526766 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctrzf\" (UniqueName: \"kubernetes.io/projected/a3fd4198-2135-4d50-8857-21055b3f99b0-kube-api-access-ctrzf\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.526776 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.538692 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-config-data" (OuterVolumeSpecName: "config-data") pod "a3fd4198-2135-4d50-8857-21055b3f99b0" (UID: "a3fd4198-2135-4d50-8857-21055b3f99b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.628858 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4198-2135-4d50-8857-21055b3f99b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.698447 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.706939 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.722846 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.723550 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="sg-core" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723580 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="sg-core" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.723597 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-central-agent" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723607 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-central-agent" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.723624 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-notification-agent" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723632 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-notification-agent" Dec 04 14:20:48 crc kubenswrapper[4715]: E1204 14:20:48.723663 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="proxy-httpd" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723671 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="proxy-httpd" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723868 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-notification-agent" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723891 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="sg-core" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723902 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="proxy-httpd" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.723910 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" containerName="ceilometer-central-agent" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.737263 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.739586 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.739801 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.760634 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833117 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-scripts\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833464 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-run-httpd\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-log-httpd\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833698 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-config-data\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833849 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.833879 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrttw\" (UniqueName: \"kubernetes.io/projected/3d563a00-f7e8-4df0-8e90-0ab639b9507c-kube-api-access-nrttw\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.935877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-scripts\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.936022 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-run-httpd\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.936098 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-log-httpd\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.936131 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.936161 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-config-data\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.936221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.936246 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrttw\" (UniqueName: \"kubernetes.io/projected/3d563a00-f7e8-4df0-8e90-0ab639b9507c-kube-api-access-nrttw\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.937427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-log-httpd\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.937885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-run-httpd\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.940990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.942576 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-scripts\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.942956 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-config-data\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.944256 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:48 crc kubenswrapper[4715]: I1204 14:20:48.961130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrttw\" (UniqueName: \"kubernetes.io/projected/3d563a00-f7e8-4df0-8e90-0ab639b9507c-kube-api-access-nrttw\") pod \"ceilometer-0\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " pod="openstack/ceilometer-0" Dec 04 14:20:49 crc kubenswrapper[4715]: I1204 14:20:49.065815 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:20:49 crc kubenswrapper[4715]: I1204 14:20:49.203920 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3fd4198-2135-4d50-8857-21055b3f99b0" path="/var/lib/kubelet/pods/a3fd4198-2135-4d50-8857-21055b3f99b0/volumes" Dec 04 14:20:49 crc kubenswrapper[4715]: I1204 14:20:49.356698 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"02266663-d1a7-4968-ab1e-e4bb6b4a9b21","Type":"ContainerStarted","Data":"1f04841fe5124fcfd0048db4e107d45f2da876c46bc8523b564ffcc8829f0025"} Dec 04 14:20:49 crc kubenswrapper[4715]: I1204 14:20:49.358730 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:49 crc kubenswrapper[4715]: I1204 14:20:49.395972 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.395940991 podStartE2EDuration="2.395940991s" podCreationTimestamp="2025-12-04 14:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:20:49.379425333 +0000 UTC m=+1406.448143538" watchObservedRunningTime="2025-12-04 14:20:49.395940991 +0000 UTC m=+1406.464659236" Dec 04 14:20:49 crc kubenswrapper[4715]: I1204 14:20:49.564891 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:20:50 crc kubenswrapper[4715]: I1204 14:20:50.370890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerStarted","Data":"1b2d02e8399091bb1b8202b31c52e69c60f9b26ff77fc7d1e318d4964937be52"} Dec 04 14:20:51 crc kubenswrapper[4715]: I1204 14:20:51.397253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerStarted","Data":"cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6"} Dec 04 14:20:51 crc kubenswrapper[4715]: I1204 14:20:51.397567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerStarted","Data":"b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d"} Dec 04 14:20:53 crc kubenswrapper[4715]: I1204 14:20:53.424236 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerStarted","Data":"27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591"} Dec 04 14:20:54 crc kubenswrapper[4715]: I1204 14:20:54.444990 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerStarted","Data":"41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60"} Dec 04 14:20:54 crc kubenswrapper[4715]: I1204 14:20:54.473190 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.169035117 podStartE2EDuration="6.47316845s" podCreationTimestamp="2025-12-04 14:20:48 +0000 UTC" firstStartedPulling="2025-12-04 14:20:49.573424737 +0000 UTC m=+1406.642142952" lastFinishedPulling="2025-12-04 14:20:53.87755807 +0000 UTC m=+1410.946276285" observedRunningTime="2025-12-04 14:20:54.467549898 +0000 UTC m=+1411.536268113" watchObservedRunningTime="2025-12-04 14:20:54.47316845 +0000 UTC m=+1411.541886665" Dec 04 14:20:55 crc kubenswrapper[4715]: I1204 14:20:55.459422 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:20:57 crc kubenswrapper[4715]: I1204 14:20:57.782171 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.317190 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2h5kl"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.319139 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.322365 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.323613 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.328784 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2h5kl"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.432007 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-scripts\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.432081 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmxtf\" (UniqueName: \"kubernetes.io/projected/54688981-4eb1-49e5-ac13-39f4f5429bab-kube-api-access-kmxtf\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.432186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-config-data\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.432304 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.533740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-scripts\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.533800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmxtf\" (UniqueName: \"kubernetes.io/projected/54688981-4eb1-49e5-ac13-39f4f5429bab-kube-api-access-kmxtf\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.533910 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-config-data\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.533973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.539980 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.541466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-scripts\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.542478 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-config-data\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.557273 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.558793 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.562941 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.569262 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmxtf\" (UniqueName: \"kubernetes.io/projected/54688981-4eb1-49e5-ac13-39f4f5429bab-kube-api-access-kmxtf\") pod \"nova-cell0-cell-mapping-2h5kl\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.583538 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.636170 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8b72\" (UniqueName: \"kubernetes.io/projected/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-kube-api-access-p8b72\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.636347 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.636421 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.644880 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.742256 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8b72\" (UniqueName: \"kubernetes.io/projected/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-kube-api-access-p8b72\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.742353 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.742445 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.754395 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.765106 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.777126 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.779028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.793135 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.799232 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.813370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8b72\" (UniqueName: \"kubernetes.io/projected/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-kube-api-access-p8b72\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.826211 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.828208 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.851607 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.853761 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-config-data\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.853867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.856333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9794g\" (UniqueName: \"kubernetes.io/projected/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-kube-api-access-9794g\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.924426 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.962874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-config-data\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.962952 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.962992 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48b0f49f-eb03-439c-84bd-076c9bdf20d7-logs\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.963075 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9794g\" (UniqueName: \"kubernetes.io/projected/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-kube-api-access-9794g\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.963158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzlw9\" (UniqueName: \"kubernetes.io/projected/48b0f49f-eb03-439c-84bd-076c9bdf20d7-kube-api-access-xzlw9\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.963202 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.963233 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-config-data\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.970734 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.979282 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-config-data\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.983748 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:20:58 crc kubenswrapper[4715]: I1204 14:20:58.997347 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9794g\" (UniqueName: \"kubernetes.io/projected/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-kube-api-access-9794g\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.005993 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.006181 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.015281 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " pod="openstack/nova-scheduler-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.039830 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.048768 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gvqf4"] Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.050868 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.075710 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzwjp\" (UniqueName: \"kubernetes.io/projected/80695692-6bc8-428b-85c8-81a0d73bed54-kube-api-access-wzwjp\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.087999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-config-data\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.088207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzlw9\" (UniqueName: \"kubernetes.io/projected/48b0f49f-eb03-439c-84bd-076c9bdf20d7-kube-api-access-xzlw9\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.088281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.088366 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80695692-6bc8-428b-85c8-81a0d73bed54-logs\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.088490 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-config-data\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.088612 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.088652 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48b0f49f-eb03-439c-84bd-076c9bdf20d7-logs\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.089172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48b0f49f-eb03-439c-84bd-076c9bdf20d7-logs\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.113204 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-config-data\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.113337 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.114593 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gvqf4"] Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.137913 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzlw9\" (UniqueName: \"kubernetes.io/projected/48b0f49f-eb03-439c-84bd-076c9bdf20d7-kube-api-access-xzlw9\") pod \"nova-metadata-0\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.184217 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.197448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.197737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.197828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq6rx\" (UniqueName: \"kubernetes.io/projected/b2a21e37-a737-4818-a332-e2e8048314c3-kube-api-access-gq6rx\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.197940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzwjp\" (UniqueName: \"kubernetes.io/projected/80695692-6bc8-428b-85c8-81a0d73bed54-kube-api-access-wzwjp\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.197994 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-config-data\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.198063 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-config\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.198496 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.198612 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.198749 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80695692-6bc8-428b-85c8-81a0d73bed54-logs\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.198897 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.203338 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80695692-6bc8-428b-85c8-81a0d73bed54-logs\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.207199 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.229898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-config-data\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.232939 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzwjp\" (UniqueName: \"kubernetes.io/projected/80695692-6bc8-428b-85c8-81a0d73bed54-kube-api-access-wzwjp\") pod \"nova-api-0\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.233478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.302588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.303209 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.303319 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq6rx\" (UniqueName: \"kubernetes.io/projected/b2a21e37-a737-4818-a332-e2e8048314c3-kube-api-access-gq6rx\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.303471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-config\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.303525 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.303629 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.305890 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.309504 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.309558 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.311313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-config\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.311475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.336246 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq6rx\" (UniqueName: \"kubernetes.io/projected/b2a21e37-a737-4818-a332-e2e8048314c3-kube-api-access-gq6rx\") pod \"dnsmasq-dns-757b4f8459-gvqf4\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.401421 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.444084 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.498004 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2h5kl"] Dec 04 14:20:59 crc kubenswrapper[4715]: W1204 14:20:59.511361 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54688981_4eb1_49e5_ac13_39f4f5429bab.slice/crio-79895751703e8c61bb98320c25e9e9c8211351798bd79ef99ba6233efb329ff3 WatchSource:0}: Error finding container 79895751703e8c61bb98320c25e9e9c8211351798bd79ef99ba6233efb329ff3: Status 404 returned error can't find the container with id 79895751703e8c61bb98320c25e9e9c8211351798bd79ef99ba6233efb329ff3 Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.789743 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.867699 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6pd76"] Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.868999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.873751 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.873822 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.880301 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6pd76"] Dec 04 14:20:59 crc kubenswrapper[4715]: I1204 14:20:59.990395 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:20:59 crc kubenswrapper[4715]: W1204 14:20:59.991700 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5071cae_6a01_4c1a_ae37_c2be194a0c1e.slice/crio-6f3f1211d4ac2c12fa236181093f1dfa0fb98b45a94417177e06a00301440705 WatchSource:0}: Error finding container 6f3f1211d4ac2c12fa236181093f1dfa0fb98b45a94417177e06a00301440705: Status 404 returned error can't find the container with id 6f3f1211d4ac2c12fa236181093f1dfa0fb98b45a94417177e06a00301440705 Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.047798 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.048232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-scripts\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.048281 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-config-data\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.048658 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclkh\" (UniqueName: \"kubernetes.io/projected/95dcc828-2041-48f7-9f22-31911926f820-kube-api-access-nclkh\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.049415 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:00 crc kubenswrapper[4715]: W1204 14:21:00.049737 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48b0f49f_eb03_439c_84bd_076c9bdf20d7.slice/crio-69d7487ff62c6fc39b4ce2e36eb47f004170861a70b7af813e74b540d391e9e6 WatchSource:0}: Error finding container 69d7487ff62c6fc39b4ce2e36eb47f004170861a70b7af813e74b540d391e9e6: Status 404 returned error can't find the container with id 69d7487ff62c6fc39b4ce2e36eb47f004170861a70b7af813e74b540d391e9e6 Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.151478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-scripts\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.151577 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-config-data\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.151644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclkh\" (UniqueName: \"kubernetes.io/projected/95dcc828-2041-48f7-9f22-31911926f820-kube-api-access-nclkh\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.151749 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.164107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-config-data\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.165885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.170533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-scripts\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.185785 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclkh\" (UniqueName: \"kubernetes.io/projected/95dcc828-2041-48f7-9f22-31911926f820-kube-api-access-nclkh\") pod \"nova-cell1-conductor-db-sync-6pd76\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.209651 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.229423 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gvqf4"] Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.245499 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.534003 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2h5kl" event={"ID":"54688981-4eb1-49e5-ac13-39f4f5429bab","Type":"ContainerStarted","Data":"79895751703e8c61bb98320c25e9e9c8211351798bd79ef99ba6233efb329ff3"} Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.536305 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a5071cae-6a01-4c1a-ae37-c2be194a0c1e","Type":"ContainerStarted","Data":"6f3f1211d4ac2c12fa236181093f1dfa0fb98b45a94417177e06a00301440705"} Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.537706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2428232-19c6-4093-bbd7-7fb0b2cbb34f","Type":"ContainerStarted","Data":"bc30ee640cf86864c1050214756c2b5b70cbe87dfe0e18f36dfbddcfc978de8f"} Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.539609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"48b0f49f-eb03-439c-84bd-076c9bdf20d7","Type":"ContainerStarted","Data":"69d7487ff62c6fc39b4ce2e36eb47f004170861a70b7af813e74b540d391e9e6"} Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.540887 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80695692-6bc8-428b-85c8-81a0d73bed54","Type":"ContainerStarted","Data":"cfe047363dab4cb91db230be2efed0988be8fb700cdb3da5a4f512776aa9ff83"} Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.547303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" event={"ID":"b2a21e37-a737-4818-a332-e2e8048314c3","Type":"ContainerStarted","Data":"0f3aa0e6b5ae02d067329fb454d2bc655eacd07b9c910d7a823814d05950d1e8"} Dec 04 14:21:00 crc kubenswrapper[4715]: I1204 14:21:00.747306 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6pd76"] Dec 04 14:21:00 crc kubenswrapper[4715]: W1204 14:21:00.807597 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95dcc828_2041_48f7_9f22_31911926f820.slice/crio-fbe0499fed1621b90006e2873cc359a8fef147e5d0a3249ae3652acfe82cd93b WatchSource:0}: Error finding container fbe0499fed1621b90006e2873cc359a8fef147e5d0a3249ae3652acfe82cd93b: Status 404 returned error can't find the container with id fbe0499fed1621b90006e2873cc359a8fef147e5d0a3249ae3652acfe82cd93b Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.568838 4715 generic.go:334] "Generic (PLEG): container finished" podID="b2a21e37-a737-4818-a332-e2e8048314c3" containerID="5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f" exitCode=0 Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.568907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" event={"ID":"b2a21e37-a737-4818-a332-e2e8048314c3","Type":"ContainerDied","Data":"5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f"} Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.574897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2h5kl" event={"ID":"54688981-4eb1-49e5-ac13-39f4f5429bab","Type":"ContainerStarted","Data":"79dc9fd13a77abbf6d0dfb63977787e93fa92bbacd9c3463ee04b10efbc4beda"} Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.578252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6pd76" event={"ID":"95dcc828-2041-48f7-9f22-31911926f820","Type":"ContainerStarted","Data":"d0f96d0cb13b050343af85e299f56a398e2ec0a27b0c24accd18dd9f10609ed1"} Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.578327 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6pd76" event={"ID":"95dcc828-2041-48f7-9f22-31911926f820","Type":"ContainerStarted","Data":"fbe0499fed1621b90006e2873cc359a8fef147e5d0a3249ae3652acfe82cd93b"} Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.626613 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6pd76" podStartSLOduration=2.626579106 podStartE2EDuration="2.626579106s" podCreationTimestamp="2025-12-04 14:20:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:01.615978749 +0000 UTC m=+1418.684696964" watchObservedRunningTime="2025-12-04 14:21:01.626579106 +0000 UTC m=+1418.695297321" Dec 04 14:21:01 crc kubenswrapper[4715]: I1204 14:21:01.644055 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2h5kl" podStartSLOduration=3.644021489 podStartE2EDuration="3.644021489s" podCreationTimestamp="2025-12-04 14:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:01.641600073 +0000 UTC m=+1418.710318308" watchObservedRunningTime="2025-12-04 14:21:01.644021489 +0000 UTC m=+1418.712739704" Dec 04 14:21:02 crc kubenswrapper[4715]: I1204 14:21:02.456097 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:21:02 crc kubenswrapper[4715]: I1204 14:21:02.469470 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.617175 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2428232-19c6-4093-bbd7-7fb0b2cbb34f","Type":"ContainerStarted","Data":"705577dcdccb35d70a196112db7107242d919fb5601ce78ce19ce8e92b5213d3"} Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.617868 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e2428232-19c6-4093-bbd7-7fb0b2cbb34f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://705577dcdccb35d70a196112db7107242d919fb5601ce78ce19ce8e92b5213d3" gracePeriod=30 Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.622975 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"48b0f49f-eb03-439c-84bd-076c9bdf20d7","Type":"ContainerStarted","Data":"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528"} Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.626498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80695692-6bc8-428b-85c8-81a0d73bed54","Type":"ContainerStarted","Data":"bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538"} Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.645458 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" event={"ID":"b2a21e37-a737-4818-a332-e2e8048314c3","Type":"ContainerStarted","Data":"18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce"} Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.646053 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.649511 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a5071cae-6a01-4c1a-ae37-c2be194a0c1e","Type":"ContainerStarted","Data":"b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd"} Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.653124 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.508700361 podStartE2EDuration="6.653100959s" podCreationTimestamp="2025-12-04 14:20:58 +0000 UTC" firstStartedPulling="2025-12-04 14:20:59.812858537 +0000 UTC m=+1416.881576752" lastFinishedPulling="2025-12-04 14:21:03.957259125 +0000 UTC m=+1421.025977350" observedRunningTime="2025-12-04 14:21:04.64169776 +0000 UTC m=+1421.710415995" watchObservedRunningTime="2025-12-04 14:21:04.653100959 +0000 UTC m=+1421.721819174" Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.680313 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" podStartSLOduration=6.680288825 podStartE2EDuration="6.680288825s" podCreationTimestamp="2025-12-04 14:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:04.670975503 +0000 UTC m=+1421.739693738" watchObservedRunningTime="2025-12-04 14:21:04.680288825 +0000 UTC m=+1421.749007040" Dec 04 14:21:04 crc kubenswrapper[4715]: I1204 14:21:04.724162 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.771731085 podStartE2EDuration="6.724137313s" podCreationTimestamp="2025-12-04 14:20:58 +0000 UTC" firstStartedPulling="2025-12-04 14:20:59.998880595 +0000 UTC m=+1417.067598810" lastFinishedPulling="2025-12-04 14:21:03.951286733 +0000 UTC m=+1421.020005038" observedRunningTime="2025-12-04 14:21:04.69154772 +0000 UTC m=+1421.760265945" watchObservedRunningTime="2025-12-04 14:21:04.724137313 +0000 UTC m=+1421.792855538" Dec 04 14:21:05 crc kubenswrapper[4715]: I1204 14:21:05.665230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"48b0f49f-eb03-439c-84bd-076c9bdf20d7","Type":"ContainerStarted","Data":"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494"} Dec 04 14:21:05 crc kubenswrapper[4715]: I1204 14:21:05.665394 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-log" containerID="cri-o://3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528" gracePeriod=30 Dec 04 14:21:05 crc kubenswrapper[4715]: I1204 14:21:05.665577 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-metadata" containerID="cri-o://d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494" gracePeriod=30 Dec 04 14:21:05 crc kubenswrapper[4715]: I1204 14:21:05.669617 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80695692-6bc8-428b-85c8-81a0d73bed54","Type":"ContainerStarted","Data":"705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730"} Dec 04 14:21:05 crc kubenswrapper[4715]: I1204 14:21:05.697994 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.792536551 podStartE2EDuration="7.697966746s" podCreationTimestamp="2025-12-04 14:20:58 +0000 UTC" firstStartedPulling="2025-12-04 14:21:00.052525748 +0000 UTC m=+1417.121243963" lastFinishedPulling="2025-12-04 14:21:03.957955943 +0000 UTC m=+1421.026674158" observedRunningTime="2025-12-04 14:21:05.687138533 +0000 UTC m=+1422.755856758" watchObservedRunningTime="2025-12-04 14:21:05.697966746 +0000 UTC m=+1422.766684961" Dec 04 14:21:05 crc kubenswrapper[4715]: I1204 14:21:05.713403 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.022233602 podStartE2EDuration="7.713380234s" podCreationTimestamp="2025-12-04 14:20:58 +0000 UTC" firstStartedPulling="2025-12-04 14:21:00.260258724 +0000 UTC m=+1417.328976949" lastFinishedPulling="2025-12-04 14:21:03.951405366 +0000 UTC m=+1421.020123581" observedRunningTime="2025-12-04 14:21:05.709887639 +0000 UTC m=+1422.778605854" watchObservedRunningTime="2025-12-04 14:21:05.713380234 +0000 UTC m=+1422.782098449" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.309615 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.402972 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-combined-ca-bundle\") pod \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.403193 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzlw9\" (UniqueName: \"kubernetes.io/projected/48b0f49f-eb03-439c-84bd-076c9bdf20d7-kube-api-access-xzlw9\") pod \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.403268 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-config-data\") pod \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.403308 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48b0f49f-eb03-439c-84bd-076c9bdf20d7-logs\") pod \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\" (UID: \"48b0f49f-eb03-439c-84bd-076c9bdf20d7\") " Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.404007 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48b0f49f-eb03-439c-84bd-076c9bdf20d7-logs" (OuterVolumeSpecName: "logs") pod "48b0f49f-eb03-439c-84bd-076c9bdf20d7" (UID: "48b0f49f-eb03-439c-84bd-076c9bdf20d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.412581 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48b0f49f-eb03-439c-84bd-076c9bdf20d7-kube-api-access-xzlw9" (OuterVolumeSpecName: "kube-api-access-xzlw9") pod "48b0f49f-eb03-439c-84bd-076c9bdf20d7" (UID: "48b0f49f-eb03-439c-84bd-076c9bdf20d7"). InnerVolumeSpecName "kube-api-access-xzlw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.443097 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48b0f49f-eb03-439c-84bd-076c9bdf20d7" (UID: "48b0f49f-eb03-439c-84bd-076c9bdf20d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.445960 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-config-data" (OuterVolumeSpecName: "config-data") pod "48b0f49f-eb03-439c-84bd-076c9bdf20d7" (UID: "48b0f49f-eb03-439c-84bd-076c9bdf20d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.505772 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzlw9\" (UniqueName: \"kubernetes.io/projected/48b0f49f-eb03-439c-84bd-076c9bdf20d7-kube-api-access-xzlw9\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.505814 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.505824 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48b0f49f-eb03-439c-84bd-076c9bdf20d7-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.505833 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48b0f49f-eb03-439c-84bd-076c9bdf20d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.687186 4715 generic.go:334] "Generic (PLEG): container finished" podID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerID="d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494" exitCode=0 Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.687232 4715 generic.go:334] "Generic (PLEG): container finished" podID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerID="3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528" exitCode=143 Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.689322 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.689782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"48b0f49f-eb03-439c-84bd-076c9bdf20d7","Type":"ContainerDied","Data":"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494"} Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.689825 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"48b0f49f-eb03-439c-84bd-076c9bdf20d7","Type":"ContainerDied","Data":"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528"} Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.689838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"48b0f49f-eb03-439c-84bd-076c9bdf20d7","Type":"ContainerDied","Data":"69d7487ff62c6fc39b4ce2e36eb47f004170861a70b7af813e74b540d391e9e6"} Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.689857 4715 scope.go:117] "RemoveContainer" containerID="d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.736258 4715 scope.go:117] "RemoveContainer" containerID="3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.737256 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.753178 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.764449 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:06 crc kubenswrapper[4715]: E1204 14:21:06.765098 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-metadata" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.765123 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-metadata" Dec 04 14:21:06 crc kubenswrapper[4715]: E1204 14:21:06.765162 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-log" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.765173 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-log" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.765370 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-log" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.765390 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" containerName="nova-metadata-metadata" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.765873 4715 scope.go:117] "RemoveContainer" containerID="d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494" Dec 04 14:21:06 crc kubenswrapper[4715]: E1204 14:21:06.766662 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494\": container with ID starting with d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494 not found: ID does not exist" containerID="d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.766711 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494"} err="failed to get container status \"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494\": rpc error: code = NotFound desc = could not find container \"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494\": container with ID starting with d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494 not found: ID does not exist" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.766734 4715 scope.go:117] "RemoveContainer" containerID="3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528" Dec 04 14:21:06 crc kubenswrapper[4715]: E1204 14:21:06.767233 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528\": container with ID starting with 3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528 not found: ID does not exist" containerID="3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.767255 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528"} err="failed to get container status \"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528\": rpc error: code = NotFound desc = could not find container \"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528\": container with ID starting with 3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528 not found: ID does not exist" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.767269 4715 scope.go:117] "RemoveContainer" containerID="d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.767439 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.767686 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494"} err="failed to get container status \"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494\": rpc error: code = NotFound desc = could not find container \"d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494\": container with ID starting with d97e4e2037dcc9808174a4d461c42b658a0e0ef70fb835c01033e9f76538a494 not found: ID does not exist" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.767744 4715 scope.go:117] "RemoveContainer" containerID="3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.768087 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528"} err="failed to get container status \"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528\": rpc error: code = NotFound desc = could not find container \"3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528\": container with ID starting with 3f6c961db668b898254fc3c2c860e0a2fc912bdeeadc720268c54a383aaa7528 not found: ID does not exist" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.769521 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.770580 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.791836 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.913187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q48c\" (UniqueName: \"kubernetes.io/projected/04f69763-03e3-43f1-9d77-5b13e87bcea5-kube-api-access-9q48c\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.913295 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-config-data\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.913338 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.913360 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:06 crc kubenswrapper[4715]: I1204 14:21:06.913415 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04f69763-03e3-43f1-9d77-5b13e87bcea5-logs\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.015400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-config-data\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.015769 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.015882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.016065 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04f69763-03e3-43f1-9d77-5b13e87bcea5-logs\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.016261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q48c\" (UniqueName: \"kubernetes.io/projected/04f69763-03e3-43f1-9d77-5b13e87bcea5-kube-api-access-9q48c\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.016629 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04f69763-03e3-43f1-9d77-5b13e87bcea5-logs\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.019857 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.020089 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.025665 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-config-data\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.036633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q48c\" (UniqueName: \"kubernetes.io/projected/04f69763-03e3-43f1-9d77-5b13e87bcea5-kube-api-access-9q48c\") pod \"nova-metadata-0\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.102364 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.246010 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48b0f49f-eb03-439c-84bd-076c9bdf20d7" path="/var/lib/kubelet/pods/48b0f49f-eb03-439c-84bd-076c9bdf20d7/volumes" Dec 04 14:21:07 crc kubenswrapper[4715]: I1204 14:21:07.694405 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:08 crc kubenswrapper[4715]: I1204 14:21:08.716343 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04f69763-03e3-43f1-9d77-5b13e87bcea5","Type":"ContainerStarted","Data":"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631"} Dec 04 14:21:08 crc kubenswrapper[4715]: I1204 14:21:08.716451 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04f69763-03e3-43f1-9d77-5b13e87bcea5","Type":"ContainerStarted","Data":"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631"} Dec 04 14:21:08 crc kubenswrapper[4715]: I1204 14:21:08.716465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04f69763-03e3-43f1-9d77-5b13e87bcea5","Type":"ContainerStarted","Data":"b7a65837562422a455d7f9b66e4094a3155c6586c49698e231a38e5dc0ec6860"} Dec 04 14:21:08 crc kubenswrapper[4715]: I1204 14:21:08.759362 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:21:08 crc kubenswrapper[4715]: I1204 14:21:08.759428 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:21:08 crc kubenswrapper[4715]: I1204 14:21:08.984672 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.195836 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.195889 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.216438 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.234637 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.234621194 podStartE2EDuration="3.234621194s" podCreationTimestamp="2025-12-04 14:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:08.742161468 +0000 UTC m=+1425.810879713" watchObservedRunningTime="2025-12-04 14:21:09.234621194 +0000 UTC m=+1426.303339409" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.403128 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.403187 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.446317 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.509426 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-swb86"] Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.509734 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="dnsmasq-dns" containerID="cri-o://c7a72ccf3f913c5adcbaad23be5746a33b02438ef6228a4f3e7dfaa38f1d10d0" gracePeriod=10 Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.736328 4715 generic.go:334] "Generic (PLEG): container finished" podID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerID="c7a72ccf3f913c5adcbaad23be5746a33b02438ef6228a4f3e7dfaa38f1d10d0" exitCode=0 Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.736527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" event={"ID":"38ade7f7-11d1-43a2-9520-f6b217314df9","Type":"ContainerDied","Data":"c7a72ccf3f913c5adcbaad23be5746a33b02438ef6228a4f3e7dfaa38f1d10d0"} Dec 04 14:21:09 crc kubenswrapper[4715]: I1204 14:21:09.787448 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.093493 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.193678 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-sb\") pod \"38ade7f7-11d1-43a2-9520-f6b217314df9\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.193759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4rtq\" (UniqueName: \"kubernetes.io/projected/38ade7f7-11d1-43a2-9520-f6b217314df9-kube-api-access-f4rtq\") pod \"38ade7f7-11d1-43a2-9520-f6b217314df9\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.193798 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-svc\") pod \"38ade7f7-11d1-43a2-9520-f6b217314df9\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.193899 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-config\") pod \"38ade7f7-11d1-43a2-9520-f6b217314df9\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.193995 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-swift-storage-0\") pod \"38ade7f7-11d1-43a2-9520-f6b217314df9\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.194051 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-nb\") pod \"38ade7f7-11d1-43a2-9520-f6b217314df9\" (UID: \"38ade7f7-11d1-43a2-9520-f6b217314df9\") " Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.202170 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ade7f7-11d1-43a2-9520-f6b217314df9-kube-api-access-f4rtq" (OuterVolumeSpecName: "kube-api-access-f4rtq") pod "38ade7f7-11d1-43a2-9520-f6b217314df9" (UID: "38ade7f7-11d1-43a2-9520-f6b217314df9"). InnerVolumeSpecName "kube-api-access-f4rtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.275655 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "38ade7f7-11d1-43a2-9520-f6b217314df9" (UID: "38ade7f7-11d1-43a2-9520-f6b217314df9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.283203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "38ade7f7-11d1-43a2-9520-f6b217314df9" (UID: "38ade7f7-11d1-43a2-9520-f6b217314df9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.299296 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "38ade7f7-11d1-43a2-9520-f6b217314df9" (UID: "38ade7f7-11d1-43a2-9520-f6b217314df9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.299649 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.299689 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4rtq\" (UniqueName: \"kubernetes.io/projected/38ade7f7-11d1-43a2-9520-f6b217314df9-kube-api-access-f4rtq\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.299709 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.299722 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.307893 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "38ade7f7-11d1-43a2-9520-f6b217314df9" (UID: "38ade7f7-11d1-43a2-9520-f6b217314df9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.316995 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-config" (OuterVolumeSpecName: "config") pod "38ade7f7-11d1-43a2-9520-f6b217314df9" (UID: "38ade7f7-11d1-43a2-9520-f6b217314df9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.401454 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.401489 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ade7f7-11d1-43a2-9520-f6b217314df9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.485325 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.485366 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.754046 4715 generic.go:334] "Generic (PLEG): container finished" podID="54688981-4eb1-49e5-ac13-39f4f5429bab" containerID="79dc9fd13a77abbf6d0dfb63977787e93fa92bbacd9c3463ee04b10efbc4beda" exitCode=0 Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.754155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2h5kl" event={"ID":"54688981-4eb1-49e5-ac13-39f4f5429bab","Type":"ContainerDied","Data":"79dc9fd13a77abbf6d0dfb63977787e93fa92bbacd9c3463ee04b10efbc4beda"} Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.759426 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" event={"ID":"38ade7f7-11d1-43a2-9520-f6b217314df9","Type":"ContainerDied","Data":"06274e7af106a34cfe4876fb1cfe115e4432fbef3916f885debf18daa295b32d"} Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.759481 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.759523 4715 scope.go:117] "RemoveContainer" containerID="c7a72ccf3f913c5adcbaad23be5746a33b02438ef6228a4f3e7dfaa38f1d10d0" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.796736 4715 scope.go:117] "RemoveContainer" containerID="2352bbb6dcce27827a0cd26ec0464d0f9b19c3beb79e8f8861956afcf3713f86" Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.825525 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-swb86"] Dec 04 14:21:10 crc kubenswrapper[4715]: I1204 14:21:10.837090 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-swb86"] Dec 04 14:21:11 crc kubenswrapper[4715]: I1204 14:21:11.192716 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" path="/var/lib/kubelet/pods/38ade7f7-11d1-43a2-9520-f6b217314df9/volumes" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.102577 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.102664 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.172858 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.244843 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-combined-ca-bundle\") pod \"54688981-4eb1-49e5-ac13-39f4f5429bab\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.245054 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-scripts\") pod \"54688981-4eb1-49e5-ac13-39f4f5429bab\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.245114 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmxtf\" (UniqueName: \"kubernetes.io/projected/54688981-4eb1-49e5-ac13-39f4f5429bab-kube-api-access-kmxtf\") pod \"54688981-4eb1-49e5-ac13-39f4f5429bab\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.245219 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-config-data\") pod \"54688981-4eb1-49e5-ac13-39f4f5429bab\" (UID: \"54688981-4eb1-49e5-ac13-39f4f5429bab\") " Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.253165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-scripts" (OuterVolumeSpecName: "scripts") pod "54688981-4eb1-49e5-ac13-39f4f5429bab" (UID: "54688981-4eb1-49e5-ac13-39f4f5429bab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.256437 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54688981-4eb1-49e5-ac13-39f4f5429bab-kube-api-access-kmxtf" (OuterVolumeSpecName: "kube-api-access-kmxtf") pod "54688981-4eb1-49e5-ac13-39f4f5429bab" (UID: "54688981-4eb1-49e5-ac13-39f4f5429bab"). InnerVolumeSpecName "kube-api-access-kmxtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.288832 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54688981-4eb1-49e5-ac13-39f4f5429bab" (UID: "54688981-4eb1-49e5-ac13-39f4f5429bab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.292792 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-config-data" (OuterVolumeSpecName: "config-data") pod "54688981-4eb1-49e5-ac13-39f4f5429bab" (UID: "54688981-4eb1-49e5-ac13-39f4f5429bab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.349069 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.349135 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.349144 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmxtf\" (UniqueName: \"kubernetes.io/projected/54688981-4eb1-49e5-ac13-39f4f5429bab-kube-api-access-kmxtf\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.349156 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54688981-4eb1-49e5-ac13-39f4f5429bab-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.798898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2h5kl" event={"ID":"54688981-4eb1-49e5-ac13-39f4f5429bab","Type":"ContainerDied","Data":"79895751703e8c61bb98320c25e9e9c8211351798bd79ef99ba6233efb329ff3"} Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.798944 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79895751703e8c61bb98320c25e9e9c8211351798bd79ef99ba6233efb329ff3" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.799321 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2h5kl" Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.957799 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.958013 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-log" containerID="cri-o://bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538" gracePeriod=30 Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.958149 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-api" containerID="cri-o://705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730" gracePeriod=30 Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.983918 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:12 crc kubenswrapper[4715]: I1204 14:21:12.984209 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" containerName="nova-scheduler-scheduler" containerID="cri-o://b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" gracePeriod=30 Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.023705 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.024629 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-log" containerID="cri-o://d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631" gracePeriod=30 Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.024682 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-metadata" containerID="cri-o://b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631" gracePeriod=30 Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.746549 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.812661 4715 generic.go:334] "Generic (PLEG): container finished" podID="80695692-6bc8-428b-85c8-81a0d73bed54" containerID="bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538" exitCode=143 Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.812750 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80695692-6bc8-428b-85c8-81a0d73bed54","Type":"ContainerDied","Data":"bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538"} Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816584 4715 generic.go:334] "Generic (PLEG): container finished" podID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerID="b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631" exitCode=0 Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816611 4715 generic.go:334] "Generic (PLEG): container finished" podID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerID="d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631" exitCode=143 Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816612 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04f69763-03e3-43f1-9d77-5b13e87bcea5","Type":"ContainerDied","Data":"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631"} Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816658 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04f69763-03e3-43f1-9d77-5b13e87bcea5","Type":"ContainerDied","Data":"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631"} Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816673 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04f69763-03e3-43f1-9d77-5b13e87bcea5","Type":"ContainerDied","Data":"b7a65837562422a455d7f9b66e4094a3155c6586c49698e231a38e5dc0ec6860"} Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816674 4715 scope.go:117] "RemoveContainer" containerID="b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.816661 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.839996 4715 scope.go:117] "RemoveContainer" containerID="d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881073 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-nova-metadata-tls-certs\") pod \"04f69763-03e3-43f1-9d77-5b13e87bcea5\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881122 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04f69763-03e3-43f1-9d77-5b13e87bcea5-logs\") pod \"04f69763-03e3-43f1-9d77-5b13e87bcea5\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881164 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-combined-ca-bundle\") pod \"04f69763-03e3-43f1-9d77-5b13e87bcea5\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881196 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q48c\" (UniqueName: \"kubernetes.io/projected/04f69763-03e3-43f1-9d77-5b13e87bcea5-kube-api-access-9q48c\") pod \"04f69763-03e3-43f1-9d77-5b13e87bcea5\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881249 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-config-data\") pod \"04f69763-03e3-43f1-9d77-5b13e87bcea5\" (UID: \"04f69763-03e3-43f1-9d77-5b13e87bcea5\") " Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04f69763-03e3-43f1-9d77-5b13e87bcea5-logs" (OuterVolumeSpecName: "logs") pod "04f69763-03e3-43f1-9d77-5b13e87bcea5" (UID: "04f69763-03e3-43f1-9d77-5b13e87bcea5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.881904 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04f69763-03e3-43f1-9d77-5b13e87bcea5-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.889122 4715 scope.go:117] "RemoveContainer" containerID="b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631" Dec 04 14:21:13 crc kubenswrapper[4715]: E1204 14:21:13.889649 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631\": container with ID starting with b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631 not found: ID does not exist" containerID="b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.889688 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631"} err="failed to get container status \"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631\": rpc error: code = NotFound desc = could not find container \"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631\": container with ID starting with b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631 not found: ID does not exist" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.889713 4715 scope.go:117] "RemoveContainer" containerID="d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631" Dec 04 14:21:13 crc kubenswrapper[4715]: E1204 14:21:13.889977 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631\": container with ID starting with d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631 not found: ID does not exist" containerID="d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.890003 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631"} err="failed to get container status \"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631\": rpc error: code = NotFound desc = could not find container \"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631\": container with ID starting with d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631 not found: ID does not exist" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.890021 4715 scope.go:117] "RemoveContainer" containerID="b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.894229 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631"} err="failed to get container status \"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631\": rpc error: code = NotFound desc = could not find container \"b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631\": container with ID starting with b871d3a5cf5ab56d7ae892ad718e34618583b42fa74e9c9f02e97c741181a631 not found: ID does not exist" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.894272 4715 scope.go:117] "RemoveContainer" containerID="d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.894549 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04f69763-03e3-43f1-9d77-5b13e87bcea5-kube-api-access-9q48c" (OuterVolumeSpecName: "kube-api-access-9q48c") pod "04f69763-03e3-43f1-9d77-5b13e87bcea5" (UID: "04f69763-03e3-43f1-9d77-5b13e87bcea5"). InnerVolumeSpecName "kube-api-access-9q48c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.897233 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631"} err="failed to get container status \"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631\": rpc error: code = NotFound desc = could not find container \"d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631\": container with ID starting with d25f119196a92f08d35c2b623039b5b09064a9b06204e6605f89fa4aecfe4631 not found: ID does not exist" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.924207 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04f69763-03e3-43f1-9d77-5b13e87bcea5" (UID: "04f69763-03e3-43f1-9d77-5b13e87bcea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.931226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-config-data" (OuterVolumeSpecName: "config-data") pod "04f69763-03e3-43f1-9d77-5b13e87bcea5" (UID: "04f69763-03e3-43f1-9d77-5b13e87bcea5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.984289 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.984347 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q48c\" (UniqueName: \"kubernetes.io/projected/04f69763-03e3-43f1-9d77-5b13e87bcea5-kube-api-access-9q48c\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:13 crc kubenswrapper[4715]: I1204 14:21:13.984386 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.035957 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "04f69763-03e3-43f1-9d77-5b13e87bcea5" (UID: "04f69763-03e3-43f1-9d77-5b13e87bcea5"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.085993 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04f69763-03e3-43f1-9d77-5b13e87bcea5-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.189216 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.192417 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.196323 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.196384 4715 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" containerName="nova-scheduler-scheduler" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.215698 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.228929 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.247413 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.248005 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54688981-4eb1-49e5-ac13-39f4f5429bab" containerName="nova-manage" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248027 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="54688981-4eb1-49e5-ac13-39f4f5429bab" containerName="nova-manage" Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.248079 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-metadata" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248087 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-metadata" Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.248103 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-log" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248112 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-log" Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.248121 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="dnsmasq-dns" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248129 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="dnsmasq-dns" Dec 04 14:21:14 crc kubenswrapper[4715]: E1204 14:21:14.248149 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="init" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248157 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="init" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248395 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="54688981-4eb1-49e5-ac13-39f4f5429bab" containerName="nova-manage" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248417 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-log" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248437 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="dnsmasq-dns" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.248449 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" containerName="nova-metadata-metadata" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.249747 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.250374 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.255500 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.255842 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.293275 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-config-data\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.293358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.293405 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shnvs\" (UniqueName: \"kubernetes.io/projected/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-kube-api-access-shnvs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.293442 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.293478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-logs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.395602 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-config-data\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.395687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.395712 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shnvs\" (UniqueName: \"kubernetes.io/projected/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-kube-api-access-shnvs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.395739 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.395775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-logs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.396323 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-logs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.399379 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.399453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-config-data\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.400761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.418949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shnvs\" (UniqueName: \"kubernetes.io/projected/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-kube-api-access-shnvs\") pod \"nova-metadata-0\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.578200 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:21:14 crc kubenswrapper[4715]: I1204 14:21:14.851169 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9776ccc5-swb86" podUID="38ade7f7-11d1-43a2-9520-f6b217314df9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: i/o timeout" Dec 04 14:21:15 crc kubenswrapper[4715]: I1204 14:21:15.090229 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:15 crc kubenswrapper[4715]: I1204 14:21:15.271508 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04f69763-03e3-43f1-9d77-5b13e87bcea5" path="/var/lib/kubelet/pods/04f69763-03e3-43f1-9d77-5b13e87bcea5/volumes" Dec 04 14:21:15 crc kubenswrapper[4715]: I1204 14:21:15.846125 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2","Type":"ContainerStarted","Data":"fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43"} Dec 04 14:21:15 crc kubenswrapper[4715]: I1204 14:21:15.846468 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2","Type":"ContainerStarted","Data":"a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9"} Dec 04 14:21:15 crc kubenswrapper[4715]: I1204 14:21:15.846482 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2","Type":"ContainerStarted","Data":"aa15f852cf93a01e088f32980b4cabb84d5fd6fa893b48194ddd254b7b3d1f73"} Dec 04 14:21:15 crc kubenswrapper[4715]: I1204 14:21:15.869154 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.869132647 podStartE2EDuration="1.869132647s" podCreationTimestamp="2025-12-04 14:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:15.86552493 +0000 UTC m=+1432.934243145" watchObservedRunningTime="2025-12-04 14:21:15.869132647 +0000 UTC m=+1432.937850882" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.600444 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.787871 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-config-data\") pod \"80695692-6bc8-428b-85c8-81a0d73bed54\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.788128 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80695692-6bc8-428b-85c8-81a0d73bed54-logs\") pod \"80695692-6bc8-428b-85c8-81a0d73bed54\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.788198 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-combined-ca-bundle\") pod \"80695692-6bc8-428b-85c8-81a0d73bed54\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.788235 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzwjp\" (UniqueName: \"kubernetes.io/projected/80695692-6bc8-428b-85c8-81a0d73bed54-kube-api-access-wzwjp\") pod \"80695692-6bc8-428b-85c8-81a0d73bed54\" (UID: \"80695692-6bc8-428b-85c8-81a0d73bed54\") " Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.788639 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80695692-6bc8-428b-85c8-81a0d73bed54-logs" (OuterVolumeSpecName: "logs") pod "80695692-6bc8-428b-85c8-81a0d73bed54" (UID: "80695692-6bc8-428b-85c8-81a0d73bed54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.793456 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80695692-6bc8-428b-85c8-81a0d73bed54-kube-api-access-wzwjp" (OuterVolumeSpecName: "kube-api-access-wzwjp") pod "80695692-6bc8-428b-85c8-81a0d73bed54" (UID: "80695692-6bc8-428b-85c8-81a0d73bed54"). InnerVolumeSpecName "kube-api-access-wzwjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.817167 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-config-data" (OuterVolumeSpecName: "config-data") pod "80695692-6bc8-428b-85c8-81a0d73bed54" (UID: "80695692-6bc8-428b-85c8-81a0d73bed54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.825281 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80695692-6bc8-428b-85c8-81a0d73bed54" (UID: "80695692-6bc8-428b-85c8-81a0d73bed54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.858297 4715 generic.go:334] "Generic (PLEG): container finished" podID="80695692-6bc8-428b-85c8-81a0d73bed54" containerID="705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730" exitCode=0 Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.858363 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.858378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80695692-6bc8-428b-85c8-81a0d73bed54","Type":"ContainerDied","Data":"705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730"} Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.858410 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80695692-6bc8-428b-85c8-81a0d73bed54","Type":"ContainerDied","Data":"cfe047363dab4cb91db230be2efed0988be8fb700cdb3da5a4f512776aa9ff83"} Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.858429 4715 scope.go:117] "RemoveContainer" containerID="705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.860993 4715 generic.go:334] "Generic (PLEG): container finished" podID="95dcc828-2041-48f7-9f22-31911926f820" containerID="d0f96d0cb13b050343af85e299f56a398e2ec0a27b0c24accd18dd9f10609ed1" exitCode=0 Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.861346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6pd76" event={"ID":"95dcc828-2041-48f7-9f22-31911926f820","Type":"ContainerDied","Data":"d0f96d0cb13b050343af85e299f56a398e2ec0a27b0c24accd18dd9f10609ed1"} Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.891757 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80695692-6bc8-428b-85c8-81a0d73bed54-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.892113 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.892144 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzwjp\" (UniqueName: \"kubernetes.io/projected/80695692-6bc8-428b-85c8-81a0d73bed54-kube-api-access-wzwjp\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.892159 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80695692-6bc8-428b-85c8-81a0d73bed54-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.975008 4715 scope.go:117] "RemoveContainer" containerID="bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538" Dec 04 14:21:16 crc kubenswrapper[4715]: I1204 14:21:16.982374 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.005342 4715 scope.go:117] "RemoveContainer" containerID="705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730" Dec 04 14:21:17 crc kubenswrapper[4715]: E1204 14:21:17.006586 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730\": container with ID starting with 705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730 not found: ID does not exist" containerID="705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.006640 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730"} err="failed to get container status \"705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730\": rpc error: code = NotFound desc = could not find container \"705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730\": container with ID starting with 705cc58fa5be0094641f70e4709acc17fe92e72d35a978380f264eaa454fd730 not found: ID does not exist" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.006671 4715 scope.go:117] "RemoveContainer" containerID="bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.008630 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: E1204 14:21:17.012883 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538\": container with ID starting with bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538 not found: ID does not exist" containerID="bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.012951 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538"} err="failed to get container status \"bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538\": rpc error: code = NotFound desc = could not find container \"bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538\": container with ID starting with bb77a332c1bacfc63ebfb54c8c312e9b4afbad9a059d9bc381d49aa25ff12538 not found: ID does not exist" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.024376 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: E1204 14:21:17.024806 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-api" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.024823 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-api" Dec 04 14:21:17 crc kubenswrapper[4715]: E1204 14:21:17.024843 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-log" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.024849 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-log" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.025106 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-log" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.025126 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" containerName="nova-api-api" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.026131 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.031923 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.043242 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.097454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.098099 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k2pm\" (UniqueName: \"kubernetes.io/projected/b69eca33-1471-4a76-be7c-88072151e8c4-kube-api-access-5k2pm\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.098343 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-config-data\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.098505 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69eca33-1471-4a76-be7c-88072151e8c4-logs\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.196415 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80695692-6bc8-428b-85c8-81a0d73bed54" path="/var/lib/kubelet/pods/80695692-6bc8-428b-85c8-81a0d73bed54/volumes" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.200536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-config-data\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.200625 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69eca33-1471-4a76-be7c-88072151e8c4-logs\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.200750 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.200821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k2pm\" (UniqueName: \"kubernetes.io/projected/b69eca33-1471-4a76-be7c-88072151e8c4-kube-api-access-5k2pm\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.212655 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69eca33-1471-4a76-be7c-88072151e8c4-logs\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.213150 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-config-data\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.218308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.240239 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k2pm\" (UniqueName: \"kubernetes.io/projected/b69eca33-1471-4a76-be7c-88072151e8c4-kube-api-access-5k2pm\") pod \"nova-api-0\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.351864 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.379232 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.404472 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-config-data\") pod \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.404599 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9794g\" (UniqueName: \"kubernetes.io/projected/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-kube-api-access-9794g\") pod \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.404639 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-combined-ca-bundle\") pod \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\" (UID: \"a5071cae-6a01-4c1a-ae37-c2be194a0c1e\") " Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.485626 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-kube-api-access-9794g" (OuterVolumeSpecName: "kube-api-access-9794g") pod "a5071cae-6a01-4c1a-ae37-c2be194a0c1e" (UID: "a5071cae-6a01-4c1a-ae37-c2be194a0c1e"). InnerVolumeSpecName "kube-api-access-9794g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.489142 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5071cae-6a01-4c1a-ae37-c2be194a0c1e" (UID: "a5071cae-6a01-4c1a-ae37-c2be194a0c1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.507204 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9794g\" (UniqueName: \"kubernetes.io/projected/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-kube-api-access-9794g\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.507241 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.513862 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-config-data" (OuterVolumeSpecName: "config-data") pod "a5071cae-6a01-4c1a-ae37-c2be194a0c1e" (UID: "a5071cae-6a01-4c1a-ae37-c2be194a0c1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.609306 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5071cae-6a01-4c1a-ae37-c2be194a0c1e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.877009 4715 generic.go:334] "Generic (PLEG): container finished" podID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" exitCode=0 Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.877091 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.877084 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a5071cae-6a01-4c1a-ae37-c2be194a0c1e","Type":"ContainerDied","Data":"b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd"} Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.877227 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a5071cae-6a01-4c1a-ae37-c2be194a0c1e","Type":"ContainerDied","Data":"6f3f1211d4ac2c12fa236181093f1dfa0fb98b45a94417177e06a00301440705"} Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.877251 4715 scope.go:117] "RemoveContainer" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.927021 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.928712 4715 scope.go:117] "RemoveContainer" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" Dec 04 14:21:17 crc kubenswrapper[4715]: E1204 14:21:17.929288 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd\": container with ID starting with b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd not found: ID does not exist" containerID="b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.929330 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd"} err="failed to get container status \"b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd\": rpc error: code = NotFound desc = could not find container \"b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd\": container with ID starting with b5f4cafdfa30d2d803fb135bb03c39c92ffe8e9e28aa232168be5c87195d1efd not found: ID does not exist" Dec 04 14:21:17 crc kubenswrapper[4715]: W1204 14:21:17.935698 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb69eca33_1471_4a76_be7c_88072151e8c4.slice/crio-c9d0a74d7dfc5037dbefed40e27343e24c42d270a033b3e20e97ed6c84c80137 WatchSource:0}: Error finding container c9d0a74d7dfc5037dbefed40e27343e24c42d270a033b3e20e97ed6c84c80137: Status 404 returned error can't find the container with id c9d0a74d7dfc5037dbefed40e27343e24c42d270a033b3e20e97ed6c84c80137 Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.938097 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.949232 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.958993 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:17 crc kubenswrapper[4715]: E1204 14:21:17.959468 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" containerName="nova-scheduler-scheduler" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.959491 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" containerName="nova-scheduler-scheduler" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.959726 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" containerName="nova-scheduler-scheduler" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.960809 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.971184 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:21:17 crc kubenswrapper[4715]: I1204 14:21:17.971877 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.023296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x86sb\" (UniqueName: \"kubernetes.io/projected/1cfe0845-21fa-452e-be72-b6321946c018-kube-api-access-x86sb\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.023401 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.023436 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-config-data\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.178236 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x86sb\" (UniqueName: \"kubernetes.io/projected/1cfe0845-21fa-452e-be72-b6321946c018-kube-api-access-x86sb\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.178312 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.178337 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-config-data\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.183807 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-config-data\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.189392 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.202651 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x86sb\" (UniqueName: \"kubernetes.io/projected/1cfe0845-21fa-452e-be72-b6321946c018-kube-api-access-x86sb\") pod \"nova-scheduler-0\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.282574 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.403001 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.585224 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-config-data\") pod \"95dcc828-2041-48f7-9f22-31911926f820\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.585348 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-combined-ca-bundle\") pod \"95dcc828-2041-48f7-9f22-31911926f820\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.585386 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nclkh\" (UniqueName: \"kubernetes.io/projected/95dcc828-2041-48f7-9f22-31911926f820-kube-api-access-nclkh\") pod \"95dcc828-2041-48f7-9f22-31911926f820\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.585669 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-scripts\") pod \"95dcc828-2041-48f7-9f22-31911926f820\" (UID: \"95dcc828-2041-48f7-9f22-31911926f820\") " Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.589757 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-scripts" (OuterVolumeSpecName: "scripts") pod "95dcc828-2041-48f7-9f22-31911926f820" (UID: "95dcc828-2041-48f7-9f22-31911926f820"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.590275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95dcc828-2041-48f7-9f22-31911926f820-kube-api-access-nclkh" (OuterVolumeSpecName: "kube-api-access-nclkh") pod "95dcc828-2041-48f7-9f22-31911926f820" (UID: "95dcc828-2041-48f7-9f22-31911926f820"). InnerVolumeSpecName "kube-api-access-nclkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.624697 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95dcc828-2041-48f7-9f22-31911926f820" (UID: "95dcc828-2041-48f7-9f22-31911926f820"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.629868 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-config-data" (OuterVolumeSpecName: "config-data") pod "95dcc828-2041-48f7-9f22-31911926f820" (UID: "95dcc828-2041-48f7-9f22-31911926f820"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.687910 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.687960 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.687976 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95dcc828-2041-48f7-9f22-31911926f820-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.687991 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nclkh\" (UniqueName: \"kubernetes.io/projected/95dcc828-2041-48f7-9f22-31911926f820-kube-api-access-nclkh\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.736941 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:18 crc kubenswrapper[4715]: W1204 14:21:18.743442 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cfe0845_21fa_452e_be72_b6321946c018.slice/crio-4e851d522a9d5f1d81b0e89a2fbca8198e71e9bf2b77fdf9d203460aeed613d3 WatchSource:0}: Error finding container 4e851d522a9d5f1d81b0e89a2fbca8198e71e9bf2b77fdf9d203460aeed613d3: Status 404 returned error can't find the container with id 4e851d522a9d5f1d81b0e89a2fbca8198e71e9bf2b77fdf9d203460aeed613d3 Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.891998 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b69eca33-1471-4a76-be7c-88072151e8c4","Type":"ContainerStarted","Data":"344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846"} Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.892089 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b69eca33-1471-4a76-be7c-88072151e8c4","Type":"ContainerStarted","Data":"62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79"} Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.892104 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b69eca33-1471-4a76-be7c-88072151e8c4","Type":"ContainerStarted","Data":"c9d0a74d7dfc5037dbefed40e27343e24c42d270a033b3e20e97ed6c84c80137"} Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.894304 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6pd76" event={"ID":"95dcc828-2041-48f7-9f22-31911926f820","Type":"ContainerDied","Data":"fbe0499fed1621b90006e2873cc359a8fef147e5d0a3249ae3652acfe82cd93b"} Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.894375 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbe0499fed1621b90006e2873cc359a8fef147e5d0a3249ae3652acfe82cd93b" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.894380 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6pd76" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.896268 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1cfe0845-21fa-452e-be72-b6321946c018","Type":"ContainerStarted","Data":"4e851d522a9d5f1d81b0e89a2fbca8198e71e9bf2b77fdf9d203460aeed613d3"} Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.922334 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.922305652 podStartE2EDuration="2.922305652s" podCreationTimestamp="2025-12-04 14:21:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:18.914798608 +0000 UTC m=+1435.983516823" watchObservedRunningTime="2025-12-04 14:21:18.922305652 +0000 UTC m=+1435.991023867" Dec 04 14:21:18 crc kubenswrapper[4715]: I1204 14:21:18.996675 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:21:19 crc kubenswrapper[4715]: E1204 14:21:19.000375 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95dcc828-2041-48f7-9f22-31911926f820" containerName="nova-cell1-conductor-db-sync" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.000417 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="95dcc828-2041-48f7-9f22-31911926f820" containerName="nova-cell1-conductor-db-sync" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.000749 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="95dcc828-2041-48f7-9f22-31911926f820" containerName="nova-cell1-conductor-db-sync" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.001645 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.004589 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.012520 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.075552 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.099910 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gfmn\" (UniqueName: \"kubernetes.io/projected/0246a68d-8dff-43c6-9efa-fef70ae0919b-kube-api-access-8gfmn\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.100061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0246a68d-8dff-43c6-9efa-fef70ae0919b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.100095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0246a68d-8dff-43c6-9efa-fef70ae0919b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.214415 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0246a68d-8dff-43c6-9efa-fef70ae0919b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.214466 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0246a68d-8dff-43c6-9efa-fef70ae0919b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.214678 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gfmn\" (UniqueName: \"kubernetes.io/projected/0246a68d-8dff-43c6-9efa-fef70ae0919b-kube-api-access-8gfmn\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.219384 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0246a68d-8dff-43c6-9efa-fef70ae0919b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.220120 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0246a68d-8dff-43c6-9efa-fef70ae0919b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.236361 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gfmn\" (UniqueName: \"kubernetes.io/projected/0246a68d-8dff-43c6-9efa-fef70ae0919b-kube-api-access-8gfmn\") pod \"nova-cell1-conductor-0\" (UID: \"0246a68d-8dff-43c6-9efa-fef70ae0919b\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.252490 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5071cae-6a01-4c1a-ae37-c2be194a0c1e" path="/var/lib/kubelet/pods/a5071cae-6a01-4c1a-ae37-c2be194a0c1e/volumes" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.324164 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.578614 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.579062 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.860835 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.913320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0246a68d-8dff-43c6-9efa-fef70ae0919b","Type":"ContainerStarted","Data":"4943f3e121ab074cbce1ff71c26e2242772032d3c9f640ca254dd13606fd5e05"} Dec 04 14:21:19 crc kubenswrapper[4715]: I1204 14:21:19.915098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1cfe0845-21fa-452e-be72-b6321946c018","Type":"ContainerStarted","Data":"fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94"} Dec 04 14:21:20 crc kubenswrapper[4715]: I1204 14:21:20.931178 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0246a68d-8dff-43c6-9efa-fef70ae0919b","Type":"ContainerStarted","Data":"7d4968e3c96e5c3f8aac9393715c4e05ebe5f50735e00a1a92ba7fa101618155"} Dec 04 14:21:20 crc kubenswrapper[4715]: I1204 14:21:20.931615 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:21 crc kubenswrapper[4715]: I1204 14:21:21.087803 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.087746415 podStartE2EDuration="4.087746415s" podCreationTimestamp="2025-12-04 14:21:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:19.939278033 +0000 UTC m=+1437.007996278" watchObservedRunningTime="2025-12-04 14:21:21.087746415 +0000 UTC m=+1438.156464630" Dec 04 14:21:21 crc kubenswrapper[4715]: I1204 14:21:21.090503 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.09049329 podStartE2EDuration="3.09049329s" podCreationTimestamp="2025-12-04 14:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:20.949680796 +0000 UTC m=+1438.018399021" watchObservedRunningTime="2025-12-04 14:21:21.09049329 +0000 UTC m=+1438.159211505" Dec 04 14:21:23 crc kubenswrapper[4715]: I1204 14:21:23.283707 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.147166 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.147699 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="435f8c50-e8e5-496f-9602-244bc202517f" containerName="kube-state-metrics" containerID="cri-o://5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b" gracePeriod=30 Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.293676 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="435f8c50-e8e5-496f-9602-244bc202517f" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": dial tcp 10.217.0.104:8081: connect: connection refused" Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.656220 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.656936 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.877433 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.964732 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkh7d\" (UniqueName: \"kubernetes.io/projected/435f8c50-e8e5-496f-9602-244bc202517f-kube-api-access-gkh7d\") pod \"435f8c50-e8e5-496f-9602-244bc202517f\" (UID: \"435f8c50-e8e5-496f-9602-244bc202517f\") " Dec 04 14:21:24 crc kubenswrapper[4715]: I1204 14:21:24.978376 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/435f8c50-e8e5-496f-9602-244bc202517f-kube-api-access-gkh7d" (OuterVolumeSpecName: "kube-api-access-gkh7d") pod "435f8c50-e8e5-496f-9602-244bc202517f" (UID: "435f8c50-e8e5-496f-9602-244bc202517f"). InnerVolumeSpecName "kube-api-access-gkh7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.061211 4715 generic.go:334] "Generic (PLEG): container finished" podID="435f8c50-e8e5-496f-9602-244bc202517f" containerID="5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b" exitCode=2 Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.061872 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435f8c50-e8e5-496f-9602-244bc202517f","Type":"ContainerDied","Data":"5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b"} Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.061935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435f8c50-e8e5-496f-9602-244bc202517f","Type":"ContainerDied","Data":"1128bc67c2442d053044f35109a2f5944f41ecb4b885ddeb750c00e53f0a3caf"} Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.061960 4715 scope.go:117] "RemoveContainer" containerID="5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.062362 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.067061 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkh7d\" (UniqueName: \"kubernetes.io/projected/435f8c50-e8e5-496f-9602-244bc202517f-kube-api-access-gkh7d\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.112768 4715 scope.go:117] "RemoveContainer" containerID="5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b" Dec 04 14:21:25 crc kubenswrapper[4715]: E1204 14:21:25.115610 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b\": container with ID starting with 5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b not found: ID does not exist" containerID="5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.115676 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b"} err="failed to get container status \"5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b\": rpc error: code = NotFound desc = could not find container \"5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b\": container with ID starting with 5fc9138f14926115cec7124af6ef8775e6fc10d1d7564fa1939fcf1d2f336d4b not found: ID does not exist" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.124785 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.157805 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.179080 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:21:25 crc kubenswrapper[4715]: E1204 14:21:25.179618 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435f8c50-e8e5-496f-9602-244bc202517f" containerName="kube-state-metrics" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.179633 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="435f8c50-e8e5-496f-9602-244bc202517f" containerName="kube-state-metrics" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.179909 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="435f8c50-e8e5-496f-9602-244bc202517f" containerName="kube-state-metrics" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.180923 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.184132 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.185785 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.382849 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="435f8c50-e8e5-496f-9602-244bc202517f" path="/var/lib/kubelet/pods/435f8c50-e8e5-496f-9602-244bc202517f/volumes" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.388179 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.484822 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.485012 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.485110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.485162 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqm4s\" (UniqueName: \"kubernetes.io/projected/b0d685d3-1f35-497c-986b-14d7407716d7-kube-api-access-xqm4s\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.586643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.586697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.586747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqm4s\" (UniqueName: \"kubernetes.io/projected/b0d685d3-1f35-497c-986b-14d7407716d7-kube-api-access-xqm4s\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.586864 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.592914 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.595605 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.596224 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0d685d3-1f35-497c-986b-14d7407716d7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.609652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqm4s\" (UniqueName: \"kubernetes.io/projected/b0d685d3-1f35-497c-986b-14d7407716d7-kube-api-access-xqm4s\") pod \"kube-state-metrics-0\" (UID: \"b0d685d3-1f35-497c-986b-14d7407716d7\") " pod="openstack/kube-state-metrics-0" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.680266 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.680291 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:25 crc kubenswrapper[4715]: I1204 14:21:25.687660 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:21:26 crc kubenswrapper[4715]: I1204 14:21:26.214959 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:21:26 crc kubenswrapper[4715]: W1204 14:21:26.217015 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0d685d3_1f35_497c_986b_14d7407716d7.slice/crio-fb6aa2d52d902163e6f5e3077b61412d17fa87d20f72764d387b57ec1befb9c3 WatchSource:0}: Error finding container fb6aa2d52d902163e6f5e3077b61412d17fa87d20f72764d387b57ec1befb9c3: Status 404 returned error can't find the container with id fb6aa2d52d902163e6f5e3077b61412d17fa87d20f72764d387b57ec1befb9c3 Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.134177 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b0d685d3-1f35-497c-986b-14d7407716d7","Type":"ContainerStarted","Data":"2690d68195b2315ec10cbd2917eea17d0462623cfd44befa806f392e24181c56"} Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.134742 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b0d685d3-1f35-497c-986b-14d7407716d7","Type":"ContainerStarted","Data":"fb6aa2d52d902163e6f5e3077b61412d17fa87d20f72764d387b57ec1befb9c3"} Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.134787 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.155449 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.777693279 podStartE2EDuration="2.155429369s" podCreationTimestamp="2025-12-04 14:21:25 +0000 UTC" firstStartedPulling="2025-12-04 14:21:26.220481709 +0000 UTC m=+1443.289199924" lastFinishedPulling="2025-12-04 14:21:26.598217799 +0000 UTC m=+1443.666936014" observedRunningTime="2025-12-04 14:21:27.148428709 +0000 UTC m=+1444.217146924" watchObservedRunningTime="2025-12-04 14:21:27.155429369 +0000 UTC m=+1444.224147584" Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.314255 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.314595 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-central-agent" containerID="cri-o://b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d" gracePeriod=30 Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.314706 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="sg-core" containerID="cri-o://27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591" gracePeriod=30 Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.314709 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="proxy-httpd" containerID="cri-o://41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60" gracePeriod=30 Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.314757 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-notification-agent" containerID="cri-o://cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6" gracePeriod=30 Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.353462 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:21:27 crc kubenswrapper[4715]: I1204 14:21:27.353510 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.146376 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerID="41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60" exitCode=0 Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.146675 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerID="27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591" exitCode=2 Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.146467 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerDied","Data":"41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60"} Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.146730 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerDied","Data":"27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591"} Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.146745 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerDied","Data":"b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d"} Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.146689 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerID="b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d" exitCode=0 Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.283674 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.314166 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.436291 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:28 crc kubenswrapper[4715]: I1204 14:21:28.436323 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.297759 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.361509 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.858110 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.984759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-scripts\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.984806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-run-httpd\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.984954 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrttw\" (UniqueName: \"kubernetes.io/projected/3d563a00-f7e8-4df0-8e90-0ab639b9507c-kube-api-access-nrttw\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.985028 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-sg-core-conf-yaml\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.985077 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-config-data\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.985099 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-log-httpd\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.985178 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-combined-ca-bundle\") pod \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\" (UID: \"3d563a00-f7e8-4df0-8e90-0ab639b9507c\") " Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.986820 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.987373 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:29 crc kubenswrapper[4715]: I1204 14:21:29.991855 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-scripts" (OuterVolumeSpecName: "scripts") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.008837 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d563a00-f7e8-4df0-8e90-0ab639b9507c-kube-api-access-nrttw" (OuterVolumeSpecName: "kube-api-access-nrttw") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "kube-api-access-nrttw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.026862 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.088924 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrttw\" (UniqueName: \"kubernetes.io/projected/3d563a00-f7e8-4df0-8e90-0ab639b9507c-kube-api-access-nrttw\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.088968 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.088984 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.088997 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.089009 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d563a00-f7e8-4df0-8e90-0ab639b9507c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.120934 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.134359 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-config-data" (OuterVolumeSpecName: "config-data") pod "3d563a00-f7e8-4df0-8e90-0ab639b9507c" (UID: "3d563a00-f7e8-4df0-8e90-0ab639b9507c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.191934 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.191971 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d563a00-f7e8-4df0-8e90-0ab639b9507c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.251413 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerID="cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6" exitCode=0 Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.252237 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.252483 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerDied","Data":"cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6"} Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.252525 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d563a00-f7e8-4df0-8e90-0ab639b9507c","Type":"ContainerDied","Data":"1b2d02e8399091bb1b8202b31c52e69c60f9b26ff77fc7d1e318d4964937be52"} Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.252550 4715 scope.go:117] "RemoveContainer" containerID="41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.312796 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.316454 4715 scope.go:117] "RemoveContainer" containerID="27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.333717 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.345865 4715 scope.go:117] "RemoveContainer" containerID="cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.346143 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.346724 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-notification-agent" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.346750 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-notification-agent" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.346779 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-central-agent" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.346788 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-central-agent" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.346804 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="proxy-httpd" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.346812 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="proxy-httpd" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.346829 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="sg-core" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.346839 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="sg-core" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.353178 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="proxy-httpd" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.353223 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-notification-agent" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.353259 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="sg-core" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.353299 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" containerName="ceilometer-central-agent" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.390484 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.413289 4715 scope.go:117] "RemoveContainer" containerID="b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.417128 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.417187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-scripts\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.421539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-log-httpd\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.420165 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.420388 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.420672 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.420994 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.421699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-config-data\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.421885 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-run-httpd\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.422121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.422177 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxvdb\" (UniqueName: \"kubernetes.io/projected/4909cf0b-6148-450d-ba35-1f28abf95901-kube-api-access-dxvdb\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.422225 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.453337 4715 scope.go:117] "RemoveContainer" containerID="41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.453920 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60\": container with ID starting with 41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60 not found: ID does not exist" containerID="41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.453961 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60"} err="failed to get container status \"41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60\": rpc error: code = NotFound desc = could not find container \"41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60\": container with ID starting with 41e886a8f462040dfe5fd088aed74f2e53dff6116940259b4beeb0ac28df9a60 not found: ID does not exist" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.453989 4715 scope.go:117] "RemoveContainer" containerID="27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.454271 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591\": container with ID starting with 27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591 not found: ID does not exist" containerID="27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.454300 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591"} err="failed to get container status \"27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591\": rpc error: code = NotFound desc = could not find container \"27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591\": container with ID starting with 27c6bb9c7d136061cc5fab8a0acbce469acf2b3463a7e1364d28c93841adc591 not found: ID does not exist" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.454318 4715 scope.go:117] "RemoveContainer" containerID="cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.455786 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6\": container with ID starting with cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6 not found: ID does not exist" containerID="cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.455831 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6"} err="failed to get container status \"cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6\": rpc error: code = NotFound desc = could not find container \"cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6\": container with ID starting with cf0e07074110af31a7c5149234ac3e429257338df82a429a838a38b31f85dae6 not found: ID does not exist" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.455856 4715 scope.go:117] "RemoveContainer" containerID="b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d" Dec 04 14:21:30 crc kubenswrapper[4715]: E1204 14:21:30.456528 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d\": container with ID starting with b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d not found: ID does not exist" containerID="b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.456584 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d"} err="failed to get container status \"b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d\": rpc error: code = NotFound desc = could not find container \"b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d\": container with ID starting with b88a5dd34fb3d2ec1e318b1ccb2ce9af1638e8a4eff63d18cbc7651c3903f23d not found: ID does not exist" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.524942 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-scripts\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.525152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-log-httpd\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.525459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-config-data\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.525686 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-run-httpd\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.525807 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.525849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxvdb\" (UniqueName: \"kubernetes.io/projected/4909cf0b-6148-450d-ba35-1f28abf95901-kube-api-access-dxvdb\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.525884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.526183 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.527899 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-run-httpd\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.528456 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-log-httpd\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.530924 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.531978 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-scripts\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.532668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.536137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.536208 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-config-data\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.548805 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxvdb\" (UniqueName: \"kubernetes.io/projected/4909cf0b-6148-450d-ba35-1f28abf95901-kube-api-access-dxvdb\") pod \"ceilometer-0\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " pod="openstack/ceilometer-0" Dec 04 14:21:30 crc kubenswrapper[4715]: I1204 14:21:30.746951 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:31 crc kubenswrapper[4715]: I1204 14:21:31.194464 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d563a00-f7e8-4df0-8e90-0ab639b9507c" path="/var/lib/kubelet/pods/3d563a00-f7e8-4df0-8e90-0ab639b9507c/volumes" Dec 04 14:21:31 crc kubenswrapper[4715]: W1204 14:21:31.290640 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4909cf0b_6148_450d_ba35_1f28abf95901.slice/crio-c80ad692862b07ea3785a771330b3e4850ff8aaf62b41ffc6a595fe64ffdf56f WatchSource:0}: Error finding container c80ad692862b07ea3785a771330b3e4850ff8aaf62b41ffc6a595fe64ffdf56f: Status 404 returned error can't find the container with id c80ad692862b07ea3785a771330b3e4850ff8aaf62b41ffc6a595fe64ffdf56f Dec 04 14:21:31 crc kubenswrapper[4715]: I1204 14:21:31.292538 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:32 crc kubenswrapper[4715]: I1204 14:21:32.287280 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerStarted","Data":"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2"} Dec 04 14:21:32 crc kubenswrapper[4715]: I1204 14:21:32.287736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerStarted","Data":"c80ad692862b07ea3785a771330b3e4850ff8aaf62b41ffc6a595fe64ffdf56f"} Dec 04 14:21:33 crc kubenswrapper[4715]: I1204 14:21:33.299591 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerStarted","Data":"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb"} Dec 04 14:21:34 crc kubenswrapper[4715]: I1204 14:21:34.312446 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerStarted","Data":"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343"} Dec 04 14:21:34 crc kubenswrapper[4715]: I1204 14:21:34.599349 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:21:34 crc kubenswrapper[4715]: I1204 14:21:34.600664 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:21:34 crc kubenswrapper[4715]: I1204 14:21:34.608344 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.328634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerStarted","Data":"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9"} Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.329270 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.332070 4715 generic.go:334] "Generic (PLEG): container finished" podID="e2428232-19c6-4093-bbd7-7fb0b2cbb34f" containerID="705577dcdccb35d70a196112db7107242d919fb5601ce78ce19ce8e92b5213d3" exitCode=137 Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.332780 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2428232-19c6-4093-bbd7-7fb0b2cbb34f","Type":"ContainerDied","Data":"705577dcdccb35d70a196112db7107242d919fb5601ce78ce19ce8e92b5213d3"} Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.341416 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.356224 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.7876958379999999 podStartE2EDuration="5.356190118s" podCreationTimestamp="2025-12-04 14:21:30 +0000 UTC" firstStartedPulling="2025-12-04 14:21:31.294388148 +0000 UTC m=+1448.363106363" lastFinishedPulling="2025-12-04 14:21:34.862882428 +0000 UTC m=+1451.931600643" observedRunningTime="2025-12-04 14:21:35.349357713 +0000 UTC m=+1452.418075948" watchObservedRunningTime="2025-12-04 14:21:35.356190118 +0000 UTC m=+1452.424908353" Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.748660 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.795574 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.991969 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8b72\" (UniqueName: \"kubernetes.io/projected/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-kube-api-access-p8b72\") pod \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.992279 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-config-data\") pod \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " Dec 04 14:21:35 crc kubenswrapper[4715]: I1204 14:21:35.992347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-combined-ca-bundle\") pod \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\" (UID: \"e2428232-19c6-4093-bbd7-7fb0b2cbb34f\") " Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.022020 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-kube-api-access-p8b72" (OuterVolumeSpecName: "kube-api-access-p8b72") pod "e2428232-19c6-4093-bbd7-7fb0b2cbb34f" (UID: "e2428232-19c6-4093-bbd7-7fb0b2cbb34f"). InnerVolumeSpecName "kube-api-access-p8b72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.046279 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2428232-19c6-4093-bbd7-7fb0b2cbb34f" (UID: "e2428232-19c6-4093-bbd7-7fb0b2cbb34f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.067544 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-config-data" (OuterVolumeSpecName: "config-data") pod "e2428232-19c6-4093-bbd7-7fb0b2cbb34f" (UID: "e2428232-19c6-4093-bbd7-7fb0b2cbb34f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.095534 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.095586 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.095601 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8b72\" (UniqueName: \"kubernetes.io/projected/e2428232-19c6-4093-bbd7-7fb0b2cbb34f-kube-api-access-p8b72\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.348626 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.350008 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2428232-19c6-4093-bbd7-7fb0b2cbb34f","Type":"ContainerDied","Data":"bc30ee640cf86864c1050214756c2b5b70cbe87dfe0e18f36dfbddcfc978de8f"} Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.350364 4715 scope.go:117] "RemoveContainer" containerID="705577dcdccb35d70a196112db7107242d919fb5601ce78ce19ce8e92b5213d3" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.415354 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.430858 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.449824 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:21:36 crc kubenswrapper[4715]: E1204 14:21:36.450358 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2428232-19c6-4093-bbd7-7fb0b2cbb34f" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.450381 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2428232-19c6-4093-bbd7-7fb0b2cbb34f" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.450619 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2428232-19c6-4093-bbd7-7fb0b2cbb34f" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.451516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.456602 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.456815 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.459305 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.463600 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.605734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-662gt\" (UniqueName: \"kubernetes.io/projected/9701ac27-1f36-4d70-a5ac-2197ea8880f6-kube-api-access-662gt\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.605832 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.606074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.606283 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.606883 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.708895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.708987 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.709069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.709119 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-662gt\" (UniqueName: \"kubernetes.io/projected/9701ac27-1f36-4d70-a5ac-2197ea8880f6-kube-api-access-662gt\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.709476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.713639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.717399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.719353 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.727259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9701ac27-1f36-4d70-a5ac-2197ea8880f6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.732744 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-662gt\" (UniqueName: \"kubernetes.io/projected/9701ac27-1f36-4d70-a5ac-2197ea8880f6-kube-api-access-662gt\") pod \"nova-cell1-novncproxy-0\" (UID: \"9701ac27-1f36-4d70-a5ac-2197ea8880f6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:36 crc kubenswrapper[4715]: I1204 14:21:36.776105 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.203949 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2428232-19c6-4093-bbd7-7fb0b2cbb34f" path="/var/lib/kubelet/pods/e2428232-19c6-4093-bbd7-7fb0b2cbb34f/volumes" Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.239925 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:21:37 crc kubenswrapper[4715]: W1204 14:21:37.255607 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9701ac27_1f36_4d70_a5ac_2197ea8880f6.slice/crio-4afac9b737fd56cf949c76f867eb6b392a2c319b33564aefe73887e908537879 WatchSource:0}: Error finding container 4afac9b737fd56cf949c76f867eb6b392a2c319b33564aefe73887e908537879: Status 404 returned error can't find the container with id 4afac9b737fd56cf949c76f867eb6b392a2c319b33564aefe73887e908537879 Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.359694 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9701ac27-1f36-4d70-a5ac-2197ea8880f6","Type":"ContainerStarted","Data":"4afac9b737fd56cf949c76f867eb6b392a2c319b33564aefe73887e908537879"} Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.364942 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.365391 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.369341 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:21:37 crc kubenswrapper[4715]: I1204 14:21:37.523281 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.381090 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9701ac27-1f36-4d70-a5ac-2197ea8880f6","Type":"ContainerStarted","Data":"f512424c579fe6721699a0349fcb5a22e90abbd66dc23eac0b9d5229ee7663d9"} Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.381642 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.385263 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.406022 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.405997182 podStartE2EDuration="2.405997182s" podCreationTimestamp="2025-12-04 14:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:38.400599726 +0000 UTC m=+1455.469317941" watchObservedRunningTime="2025-12-04 14:21:38.405997182 +0000 UTC m=+1455.474715397" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.592400 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4wwbd"] Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.594593 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.626272 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4wwbd"] Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.659271 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.659359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5dfs\" (UniqueName: \"kubernetes.io/projected/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-kube-api-access-q5dfs\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.659399 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.659427 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.659457 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.659485 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-config\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.762274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5dfs\" (UniqueName: \"kubernetes.io/projected/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-kube-api-access-q5dfs\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.762356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.762395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.762436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.762471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-config\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.762572 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.763672 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.764298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.767729 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.767868 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.768030 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-config\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.768511 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.768556 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.792812 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5dfs\" (UniqueName: \"kubernetes.io/projected/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-kube-api-access-q5dfs\") pod \"dnsmasq-dns-89c5cd4d5-4wwbd\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:38 crc kubenswrapper[4715]: I1204 14:21:38.930254 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:39 crc kubenswrapper[4715]: I1204 14:21:39.600763 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4wwbd"] Dec 04 14:21:40 crc kubenswrapper[4715]: I1204 14:21:40.407018 4715 generic.go:334] "Generic (PLEG): container finished" podID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerID="39feee6dbda3f2962e579ed067fb5942d8ec558e88cd78dfba25851f88627f5c" exitCode=0 Dec 04 14:21:40 crc kubenswrapper[4715]: I1204 14:21:40.407099 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" event={"ID":"8e7ccc35-38ad-4677-9cf3-305bc0d322c4","Type":"ContainerDied","Data":"39feee6dbda3f2962e579ed067fb5942d8ec558e88cd78dfba25851f88627f5c"} Dec 04 14:21:40 crc kubenswrapper[4715]: I1204 14:21:40.407421 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" event={"ID":"8e7ccc35-38ad-4677-9cf3-305bc0d322c4","Type":"ContainerStarted","Data":"04252d6a21d8ff9855a3c8e22aad6ce308b5fd0885e810797e14ce80786fe7a4"} Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.420187 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" event={"ID":"8e7ccc35-38ad-4677-9cf3-305bc0d322c4","Type":"ContainerStarted","Data":"5c2ecfd532559c3fe36e2edcdb4dc3f60982356d9bd62a904f2e04d27f78619f"} Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.420702 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.474180 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" podStartSLOduration=3.4741550930000002 podStartE2EDuration="3.474155093s" podCreationTimestamp="2025-12-04 14:21:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:41.447900102 +0000 UTC m=+1458.516618337" watchObservedRunningTime="2025-12-04 14:21:41.474155093 +0000 UTC m=+1458.542873318" Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.479712 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.479962 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-log" containerID="cri-o://62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79" gracePeriod=30 Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.480578 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-api" containerID="cri-o://344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846" gracePeriod=30 Dec 04 14:21:41 crc kubenswrapper[4715]: I1204 14:21:41.777363 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.431482 4715 generic.go:334] "Generic (PLEG): container finished" podID="b69eca33-1471-4a76-be7c-88072151e8c4" containerID="62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79" exitCode=143 Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.431574 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b69eca33-1471-4a76-be7c-88072151e8c4","Type":"ContainerDied","Data":"62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79"} Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.542396 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.542773 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-central-agent" containerID="cri-o://4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" gracePeriod=30 Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.542966 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="proxy-httpd" containerID="cri-o://98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" gracePeriod=30 Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.543069 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="sg-core" containerID="cri-o://9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" gracePeriod=30 Dec 04 14:21:42 crc kubenswrapper[4715]: I1204 14:21:42.543942 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-notification-agent" containerID="cri-o://de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" gracePeriod=30 Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.337053 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444609 4715 generic.go:334] "Generic (PLEG): container finished" podID="4909cf0b-6148-450d-ba35-1f28abf95901" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" exitCode=0 Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444645 4715 generic.go:334] "Generic (PLEG): container finished" podID="4909cf0b-6148-450d-ba35-1f28abf95901" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" exitCode=2 Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444652 4715 generic.go:334] "Generic (PLEG): container finished" podID="4909cf0b-6148-450d-ba35-1f28abf95901" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" exitCode=0 Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444662 4715 generic.go:334] "Generic (PLEG): container finished" podID="4909cf0b-6148-450d-ba35-1f28abf95901" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" exitCode=0 Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerDied","Data":"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9"} Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444692 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerDied","Data":"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343"} Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444755 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerDied","Data":"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb"} Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444797 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerDied","Data":"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2"} Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444807 4715 scope.go:117] "RemoveContainer" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.444813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4909cf0b-6148-450d-ba35-1f28abf95901","Type":"ContainerDied","Data":"c80ad692862b07ea3785a771330b3e4850ff8aaf62b41ffc6a595fe64ffdf56f"} Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.468814 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-config-data\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.468880 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxvdb\" (UniqueName: \"kubernetes.io/projected/4909cf0b-6148-450d-ba35-1f28abf95901-kube-api-access-dxvdb\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.468995 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-log-httpd\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.469076 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-scripts\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.469133 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-run-httpd\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.469163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-sg-core-conf-yaml\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.469210 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-ceilometer-tls-certs\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.469294 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-combined-ca-bundle\") pod \"4909cf0b-6148-450d-ba35-1f28abf95901\" (UID: \"4909cf0b-6148-450d-ba35-1f28abf95901\") " Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.469555 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.470143 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.470561 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.470815 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4909cf0b-6148-450d-ba35-1f28abf95901-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.475483 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4909cf0b-6148-450d-ba35-1f28abf95901-kube-api-access-dxvdb" (OuterVolumeSpecName: "kube-api-access-dxvdb") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "kube-api-access-dxvdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.476350 4715 scope.go:117] "RemoveContainer" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.477173 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-scripts" (OuterVolumeSpecName: "scripts") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.510597 4715 scope.go:117] "RemoveContainer" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.511278 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.529746 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.538788 4715 scope.go:117] "RemoveContainer" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.555420 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.562496 4715 scope.go:117] "RemoveContainer" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.563164 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": container with ID starting with 98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9 not found: ID does not exist" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.563223 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9"} err="failed to get container status \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": rpc error: code = NotFound desc = could not find container \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": container with ID starting with 98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.563255 4715 scope.go:117] "RemoveContainer" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.564024 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": container with ID starting with 9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343 not found: ID does not exist" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.564073 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343"} err="failed to get container status \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": rpc error: code = NotFound desc = could not find container \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": container with ID starting with 9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.564087 4715 scope.go:117] "RemoveContainer" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.564412 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": container with ID starting with de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb not found: ID does not exist" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.564435 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb"} err="failed to get container status \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": rpc error: code = NotFound desc = could not find container \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": container with ID starting with de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.564450 4715 scope.go:117] "RemoveContainer" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.565396 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": container with ID starting with 4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2 not found: ID does not exist" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.565508 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2"} err="failed to get container status \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": rpc error: code = NotFound desc = could not find container \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": container with ID starting with 4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.565604 4715 scope.go:117] "RemoveContainer" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566010 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9"} err="failed to get container status \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": rpc error: code = NotFound desc = could not find container \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": container with ID starting with 98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566057 4715 scope.go:117] "RemoveContainer" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566287 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343"} err="failed to get container status \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": rpc error: code = NotFound desc = could not find container \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": container with ID starting with 9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566320 4715 scope.go:117] "RemoveContainer" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566578 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb"} err="failed to get container status \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": rpc error: code = NotFound desc = could not find container \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": container with ID starting with de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566602 4715 scope.go:117] "RemoveContainer" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566803 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2"} err="failed to get container status \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": rpc error: code = NotFound desc = could not find container \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": container with ID starting with 4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.566823 4715 scope.go:117] "RemoveContainer" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567136 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9"} err="failed to get container status \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": rpc error: code = NotFound desc = could not find container \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": container with ID starting with 98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567155 4715 scope.go:117] "RemoveContainer" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567375 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343"} err="failed to get container status \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": rpc error: code = NotFound desc = could not find container \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": container with ID starting with 9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567393 4715 scope.go:117] "RemoveContainer" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567563 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb"} err="failed to get container status \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": rpc error: code = NotFound desc = could not find container \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": container with ID starting with de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567582 4715 scope.go:117] "RemoveContainer" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567844 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2"} err="failed to get container status \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": rpc error: code = NotFound desc = could not find container \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": container with ID starting with 4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.567868 4715 scope.go:117] "RemoveContainer" containerID="98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568125 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9"} err="failed to get container status \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": rpc error: code = NotFound desc = could not find container \"98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9\": container with ID starting with 98516caeea0ad22bf4f32af0701964601eebaea9cdb484af8629ed06fc5467a9 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568150 4715 scope.go:117] "RemoveContainer" containerID="9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568415 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343"} err="failed to get container status \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": rpc error: code = NotFound desc = could not find container \"9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343\": container with ID starting with 9b7067e340d3d1693ddf2bcd74b2c2b3f8b3288f094ad8f656c2b2ffbdff9343 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568436 4715 scope.go:117] "RemoveContainer" containerID="de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568672 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb"} err="failed to get container status \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": rpc error: code = NotFound desc = could not find container \"de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb\": container with ID starting with de59927752d3c6a2553ff73bf2f2ad88f3b7391e59ecdd6ef6ecd017bcc4a5fb not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568696 4715 scope.go:117] "RemoveContainer" containerID="4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.568960 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2"} err="failed to get container status \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": rpc error: code = NotFound desc = could not find container \"4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2\": container with ID starting with 4fe48aa13ad82bdc065f1025e02e38aff5ccc51c3a95fdf99dbec0b5571771f2 not found: ID does not exist" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.572975 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.573015 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxvdb\" (UniqueName: \"kubernetes.io/projected/4909cf0b-6148-450d-ba35-1f28abf95901-kube-api-access-dxvdb\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.573046 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.573059 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.573070 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.604742 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-config-data" (OuterVolumeSpecName: "config-data") pod "4909cf0b-6148-450d-ba35-1f28abf95901" (UID: "4909cf0b-6148-450d-ba35-1f28abf95901"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.674934 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4909cf0b-6148-450d-ba35-1f28abf95901-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.816808 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.846733 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.855692 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.856278 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-notification-agent" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856304 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-notification-agent" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.856324 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="proxy-httpd" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856331 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="proxy-httpd" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.856352 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="sg-core" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856360 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="sg-core" Dec 04 14:21:43 crc kubenswrapper[4715]: E1204 14:21:43.856398 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-central-agent" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856407 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-central-agent" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856743 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-central-agent" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856773 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="ceilometer-notification-agent" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856800 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="proxy-httpd" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.856814 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" containerName="sg-core" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.859069 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.861326 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.861549 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.861690 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.873202 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.982203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.982328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-scripts\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.982482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-config-data\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.982566 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-log-httpd\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.982618 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.983519 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.983626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-kube-api-access-zqnc7\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:43 crc kubenswrapper[4715]: I1204 14:21:43.983797 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-run-httpd\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127632 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127686 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-scripts\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127738 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-config-data\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127784 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-log-httpd\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127816 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-kube-api-access-zqnc7\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127906 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.127958 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-run-httpd\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.128519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-run-httpd\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.130046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-log-httpd\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.132572 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.133430 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.133483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-scripts\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.134178 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-config-data\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.134266 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.161844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-kube-api-access-zqnc7\") pod \"ceilometer-0\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.175715 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.611366 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:44 crc kubenswrapper[4715]: I1204 14:21:44.646878 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.193342 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4909cf0b-6148-450d-ba35-1f28abf95901" path="/var/lib/kubelet/pods/4909cf0b-6148-450d-ba35-1f28abf95901/volumes" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.324279 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.460564 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k2pm\" (UniqueName: \"kubernetes.io/projected/b69eca33-1471-4a76-be7c-88072151e8c4-kube-api-access-5k2pm\") pod \"b69eca33-1471-4a76-be7c-88072151e8c4\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.460646 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-config-data\") pod \"b69eca33-1471-4a76-be7c-88072151e8c4\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.462116 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-combined-ca-bundle\") pod \"b69eca33-1471-4a76-be7c-88072151e8c4\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.462162 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69eca33-1471-4a76-be7c-88072151e8c4-logs\") pod \"b69eca33-1471-4a76-be7c-88072151e8c4\" (UID: \"b69eca33-1471-4a76-be7c-88072151e8c4\") " Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.463626 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b69eca33-1471-4a76-be7c-88072151e8c4-logs" (OuterVolumeSpecName: "logs") pod "b69eca33-1471-4a76-be7c-88072151e8c4" (UID: "b69eca33-1471-4a76-be7c-88072151e8c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.471342 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b69eca33-1471-4a76-be7c-88072151e8c4-kube-api-access-5k2pm" (OuterVolumeSpecName: "kube-api-access-5k2pm") pod "b69eca33-1471-4a76-be7c-88072151e8c4" (UID: "b69eca33-1471-4a76-be7c-88072151e8c4"). InnerVolumeSpecName "kube-api-access-5k2pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.477319 4715 generic.go:334] "Generic (PLEG): container finished" podID="b69eca33-1471-4a76-be7c-88072151e8c4" containerID="344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846" exitCode=0 Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.477413 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.477409 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b69eca33-1471-4a76-be7c-88072151e8c4","Type":"ContainerDied","Data":"344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846"} Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.477461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b69eca33-1471-4a76-be7c-88072151e8c4","Type":"ContainerDied","Data":"c9d0a74d7dfc5037dbefed40e27343e24c42d270a033b3e20e97ed6c84c80137"} Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.477478 4715 scope.go:117] "RemoveContainer" containerID="344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.481195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerStarted","Data":"fe72778d7b45f0029843823c44361192abdc39d79a4cb46474ac5699d263c1b4"} Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.505665 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b69eca33-1471-4a76-be7c-88072151e8c4" (UID: "b69eca33-1471-4a76-be7c-88072151e8c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.525256 4715 scope.go:117] "RemoveContainer" containerID="62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.536209 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-config-data" (OuterVolumeSpecName: "config-data") pod "b69eca33-1471-4a76-be7c-88072151e8c4" (UID: "b69eca33-1471-4a76-be7c-88072151e8c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.557165 4715 scope.go:117] "RemoveContainer" containerID="344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846" Dec 04 14:21:45 crc kubenswrapper[4715]: E1204 14:21:45.557824 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846\": container with ID starting with 344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846 not found: ID does not exist" containerID="344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.557865 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846"} err="failed to get container status \"344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846\": rpc error: code = NotFound desc = could not find container \"344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846\": container with ID starting with 344e308399a71594616f59e144de4d4f8fa63cb9452714a2802b801561997846 not found: ID does not exist" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.557894 4715 scope.go:117] "RemoveContainer" containerID="62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79" Dec 04 14:21:45 crc kubenswrapper[4715]: E1204 14:21:45.558397 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79\": container with ID starting with 62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79 not found: ID does not exist" containerID="62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.558419 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79"} err="failed to get container status \"62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79\": rpc error: code = NotFound desc = could not find container \"62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79\": container with ID starting with 62520c554a6b99d2b7d3b1a23c4de7cf69ef70316bbaf2d63d7b12452b147f79 not found: ID does not exist" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.565182 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.565231 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69eca33-1471-4a76-be7c-88072151e8c4-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.565247 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k2pm\" (UniqueName: \"kubernetes.io/projected/b69eca33-1471-4a76-be7c-88072151e8c4-kube-api-access-5k2pm\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.565802 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69eca33-1471-4a76-be7c-88072151e8c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.828824 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.842818 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.859986 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:45 crc kubenswrapper[4715]: E1204 14:21:45.860488 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-log" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.860512 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-log" Dec 04 14:21:45 crc kubenswrapper[4715]: E1204 14:21:45.860550 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-api" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.860558 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-api" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.860771 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-api" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.860813 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" containerName="nova-api-log" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.866774 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.872793 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.874463 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.880308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.880665 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-config-data\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.880742 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-public-tls-certs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.880969 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.881098 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlpst\" (UniqueName: \"kubernetes.io/projected/7a28e21c-f41b-4f6e-b2f5-303bc0828613-kube-api-access-dlpst\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.881207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a28e21c-f41b-4f6e-b2f5-303bc0828613-logs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.887884 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.888791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.984228 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.984312 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlpst\" (UniqueName: \"kubernetes.io/projected/7a28e21c-f41b-4f6e-b2f5-303bc0828613-kube-api-access-dlpst\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.984348 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a28e21c-f41b-4f6e-b2f5-303bc0828613-logs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.984395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.984503 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-config-data\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.984549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-public-tls-certs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.985621 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a28e21c-f41b-4f6e-b2f5-303bc0828613-logs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.992185 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:45 crc kubenswrapper[4715]: I1204 14:21:45.997700 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-public-tls-certs\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:45.999575 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-config-data\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.005653 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.020210 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlpst\" (UniqueName: \"kubernetes.io/projected/7a28e21c-f41b-4f6e-b2f5-303bc0828613-kube-api-access-dlpst\") pod \"nova-api-0\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " pod="openstack/nova-api-0" Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.192757 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.504000 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerStarted","Data":"545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6"} Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.504721 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerStarted","Data":"c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff"} Dec 04 14:21:46 crc kubenswrapper[4715]: W1204 14:21:46.751855 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a28e21c_f41b_4f6e_b2f5_303bc0828613.slice/crio-328b95b80dc8687b515c93c4eb3bb34f83d25d6abbaeddaf37c411e954463c3c WatchSource:0}: Error finding container 328b95b80dc8687b515c93c4eb3bb34f83d25d6abbaeddaf37c411e954463c3c: Status 404 returned error can't find the container with id 328b95b80dc8687b515c93c4eb3bb34f83d25d6abbaeddaf37c411e954463c3c Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.751937 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.777409 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:46 crc kubenswrapper[4715]: I1204 14:21:46.799393 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.202700 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b69eca33-1471-4a76-be7c-88072151e8c4" path="/var/lib/kubelet/pods/b69eca33-1471-4a76-be7c-88072151e8c4/volumes" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.517254 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a28e21c-f41b-4f6e-b2f5-303bc0828613","Type":"ContainerStarted","Data":"90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771"} Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.517308 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a28e21c-f41b-4f6e-b2f5-303bc0828613","Type":"ContainerStarted","Data":"b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4"} Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.517319 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a28e21c-f41b-4f6e-b2f5-303bc0828613","Type":"ContainerStarted","Data":"328b95b80dc8687b515c93c4eb3bb34f83d25d6abbaeddaf37c411e954463c3c"} Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.525126 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerStarted","Data":"2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0"} Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.542939 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.542916454 podStartE2EDuration="2.542916454s" podCreationTimestamp="2025-12-04 14:21:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:47.539685096 +0000 UTC m=+1464.608403321" watchObservedRunningTime="2025-12-04 14:21:47.542916454 +0000 UTC m=+1464.611634669" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.545906 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.720219 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-7hlzp"] Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.733118 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7hlzp"] Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.733223 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.738798 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.738947 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.831590 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.831656 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-config-data\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.831711 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-scripts\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.831866 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrbm7\" (UniqueName: \"kubernetes.io/projected/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-kube-api-access-rrbm7\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.936533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-scripts\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.937239 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrbm7\" (UniqueName: \"kubernetes.io/projected/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-kube-api-access-rrbm7\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.937331 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.937435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-config-data\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.943242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.943890 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-config-data\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.947923 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-scripts\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:47 crc kubenswrapper[4715]: I1204 14:21:47.966853 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrbm7\" (UniqueName: \"kubernetes.io/projected/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-kube-api-access-rrbm7\") pod \"nova-cell1-cell-mapping-7hlzp\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.095094 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.547589 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-central-agent" containerID="cri-o://c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff" gracePeriod=30 Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.548155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerStarted","Data":"fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d"} Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.548655 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.548918 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="proxy-httpd" containerID="cri-o://fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d" gracePeriod=30 Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.548974 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="sg-core" containerID="cri-o://2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0" gracePeriod=30 Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.549011 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-notification-agent" containerID="cri-o://545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6" gracePeriod=30 Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.564431 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7hlzp"] Dec 04 14:21:48 crc kubenswrapper[4715]: W1204 14:21:48.572074 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87e3b168_dd77_4ba1_bdd4_dd4072b30dfb.slice/crio-2f763d4f0c10fb4d5fcb877c5b73a5188cda55b751cc2ee78049b85cd5f05b47 WatchSource:0}: Error finding container 2f763d4f0c10fb4d5fcb877c5b73a5188cda55b751cc2ee78049b85cd5f05b47: Status 404 returned error can't find the container with id 2f763d4f0c10fb4d5fcb877c5b73a5188cda55b751cc2ee78049b85cd5f05b47 Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.580953 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.334945939 podStartE2EDuration="5.580927095s" podCreationTimestamp="2025-12-04 14:21:43 +0000 UTC" firstStartedPulling="2025-12-04 14:21:44.655676473 +0000 UTC m=+1461.724394688" lastFinishedPulling="2025-12-04 14:21:47.901657629 +0000 UTC m=+1464.970375844" observedRunningTime="2025-12-04 14:21:48.57113962 +0000 UTC m=+1465.639857855" watchObservedRunningTime="2025-12-04 14:21:48.580927095 +0000 UTC m=+1465.649645320" Dec 04 14:21:48 crc kubenswrapper[4715]: I1204 14:21:48.931257 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.011708 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gvqf4"] Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.013233 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="dnsmasq-dns" containerID="cri-o://18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce" gracePeriod=10 Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.529934 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.567730 4715 generic.go:334] "Generic (PLEG): container finished" podID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerID="fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d" exitCode=0 Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.567764 4715 generic.go:334] "Generic (PLEG): container finished" podID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerID="2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0" exitCode=2 Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.567797 4715 generic.go:334] "Generic (PLEG): container finished" podID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerID="545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6" exitCode=0 Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.567844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerDied","Data":"fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.567897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerDied","Data":"2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.567910 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerDied","Data":"545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.570115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7hlzp" event={"ID":"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb","Type":"ContainerStarted","Data":"e624bc6b5d84e244abb219db0871ef091414caa93421ce28f0ce179a22af2a24"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.570142 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7hlzp" event={"ID":"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb","Type":"ContainerStarted","Data":"2f763d4f0c10fb4d5fcb877c5b73a5188cda55b751cc2ee78049b85cd5f05b47"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.572587 4715 generic.go:334] "Generic (PLEG): container finished" podID="b2a21e37-a737-4818-a332-e2e8048314c3" containerID="18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce" exitCode=0 Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.572630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" event={"ID":"b2a21e37-a737-4818-a332-e2e8048314c3","Type":"ContainerDied","Data":"18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.572681 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" event={"ID":"b2a21e37-a737-4818-a332-e2e8048314c3","Type":"ContainerDied","Data":"0f3aa0e6b5ae02d067329fb454d2bc655eacd07b9c910d7a823814d05950d1e8"} Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.572702 4715 scope.go:117] "RemoveContainer" containerID="18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.572642 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.574751 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-sb\") pod \"b2a21e37-a737-4818-a332-e2e8048314c3\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.574865 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-swift-storage-0\") pod \"b2a21e37-a737-4818-a332-e2e8048314c3\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.574889 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-nb\") pod \"b2a21e37-a737-4818-a332-e2e8048314c3\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.574943 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq6rx\" (UniqueName: \"kubernetes.io/projected/b2a21e37-a737-4818-a332-e2e8048314c3-kube-api-access-gq6rx\") pod \"b2a21e37-a737-4818-a332-e2e8048314c3\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.575077 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-svc\") pod \"b2a21e37-a737-4818-a332-e2e8048314c3\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.575114 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-config\") pod \"b2a21e37-a737-4818-a332-e2e8048314c3\" (UID: \"b2a21e37-a737-4818-a332-e2e8048314c3\") " Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.581777 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a21e37-a737-4818-a332-e2e8048314c3-kube-api-access-gq6rx" (OuterVolumeSpecName: "kube-api-access-gq6rx") pod "b2a21e37-a737-4818-a332-e2e8048314c3" (UID: "b2a21e37-a737-4818-a332-e2e8048314c3"). InnerVolumeSpecName "kube-api-access-gq6rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.598232 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-7hlzp" podStartSLOduration=2.5982150539999997 podStartE2EDuration="2.598215054s" podCreationTimestamp="2025-12-04 14:21:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:21:49.590195207 +0000 UTC m=+1466.658913422" watchObservedRunningTime="2025-12-04 14:21:49.598215054 +0000 UTC m=+1466.666933269" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.618165 4715 scope.go:117] "RemoveContainer" containerID="5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.647401 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2a21e37-a737-4818-a332-e2e8048314c3" (UID: "b2a21e37-a737-4818-a332-e2e8048314c3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.648087 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2a21e37-a737-4818-a332-e2e8048314c3" (UID: "b2a21e37-a737-4818-a332-e2e8048314c3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.655546 4715 scope.go:117] "RemoveContainer" containerID="18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.656462 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b2a21e37-a737-4818-a332-e2e8048314c3" (UID: "b2a21e37-a737-4818-a332-e2e8048314c3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:49 crc kubenswrapper[4715]: E1204 14:21:49.657467 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce\": container with ID starting with 18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce not found: ID does not exist" containerID="18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.657513 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce"} err="failed to get container status \"18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce\": rpc error: code = NotFound desc = could not find container \"18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce\": container with ID starting with 18b2cde7fb96ac0df19925e5517675dbe4beba187252cc0288b7aa725056d4ce not found: ID does not exist" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.657542 4715 scope.go:117] "RemoveContainer" containerID="5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.658069 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-config" (OuterVolumeSpecName: "config") pod "b2a21e37-a737-4818-a332-e2e8048314c3" (UID: "b2a21e37-a737-4818-a332-e2e8048314c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:49 crc kubenswrapper[4715]: E1204 14:21:49.658348 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f\": container with ID starting with 5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f not found: ID does not exist" containerID="5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.658383 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f"} err="failed to get container status \"5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f\": rpc error: code = NotFound desc = could not find container \"5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f\": container with ID starting with 5a6a7312d6518b81a9b8b98f76bcd6b8b1555023d97f9ba6ff124a220f40a82f not found: ID does not exist" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.675018 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2a21e37-a737-4818-a332-e2e8048314c3" (UID: "b2a21e37-a737-4818-a332-e2e8048314c3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.677365 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.677405 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.677421 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq6rx\" (UniqueName: \"kubernetes.io/projected/b2a21e37-a737-4818-a332-e2e8048314c3-kube-api-access-gq6rx\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.677438 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.677449 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.677462 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2a21e37-a737-4818-a332-e2e8048314c3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.909524 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gvqf4"] Dec 04 14:21:49 crc kubenswrapper[4715]: I1204 14:21:49.919797 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gvqf4"] Dec 04 14:21:51 crc kubenswrapper[4715]: I1204 14:21:51.197394 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" path="/var/lib/kubelet/pods/b2a21e37-a737-4818-a332-e2e8048314c3/volumes" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.483593 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.611964 4715 generic.go:334] "Generic (PLEG): container finished" podID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerID="c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff" exitCode=0 Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.612144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerDied","Data":"c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff"} Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.612250 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.612464 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b71c3c74-923f-4fe9-ad3b-b74e3ce55447","Type":"ContainerDied","Data":"fe72778d7b45f0029843823c44361192abdc39d79a4cb46474ac5699d263c1b4"} Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.612557 4715 scope.go:117] "RemoveContainer" containerID="fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630671 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-config-data\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630717 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-run-httpd\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-sg-core-conf-yaml\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630801 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-combined-ca-bundle\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630818 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-ceilometer-tls-certs\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630882 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-log-httpd\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-scripts\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.630993 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-kube-api-access-zqnc7\") pod \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\" (UID: \"b71c3c74-923f-4fe9-ad3b-b74e3ce55447\") " Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.631268 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.631640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.632200 4715 scope.go:117] "RemoveContainer" containerID="2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.632678 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.632702 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.641286 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-kube-api-access-zqnc7" (OuterVolumeSpecName: "kube-api-access-zqnc7") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "kube-api-access-zqnc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.659468 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.662385 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-scripts" (OuterVolumeSpecName: "scripts") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.682444 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.729653 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.734017 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-kube-api-access-zqnc7\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.734060 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.734073 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.734120 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.734137 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.745457 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-config-data" (OuterVolumeSpecName: "config-data") pod "b71c3c74-923f-4fe9-ad3b-b74e3ce55447" (UID: "b71c3c74-923f-4fe9-ad3b-b74e3ce55447"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.784719 4715 scope.go:117] "RemoveContainer" containerID="545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.804970 4715 scope.go:117] "RemoveContainer" containerID="c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.824304 4715 scope.go:117] "RemoveContainer" containerID="fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.824706 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d\": container with ID starting with fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d not found: ID does not exist" containerID="fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.824738 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d"} err="failed to get container status \"fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d\": rpc error: code = NotFound desc = could not find container \"fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d\": container with ID starting with fec8a6a43f4c2d51adaf6307e3dbc6c629de4878aaa3d5dff7d64e9fdf0e037d not found: ID does not exist" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.824762 4715 scope.go:117] "RemoveContainer" containerID="2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.825134 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0\": container with ID starting with 2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0 not found: ID does not exist" containerID="2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.825167 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0"} err="failed to get container status \"2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0\": rpc error: code = NotFound desc = could not find container \"2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0\": container with ID starting with 2bbfb16433b870aadc9600eed1d760d60357e19371aa3e1e78e1789c5243bde0 not found: ID does not exist" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.825183 4715 scope.go:117] "RemoveContainer" containerID="545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.825411 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6\": container with ID starting with 545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6 not found: ID does not exist" containerID="545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.825432 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6"} err="failed to get container status \"545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6\": rpc error: code = NotFound desc = could not find container \"545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6\": container with ID starting with 545cb4e1da44e4eaa1c15e7460f2f357c131bb28f8e2524f0c217d8ff6726bb6 not found: ID does not exist" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.825443 4715 scope.go:117] "RemoveContainer" containerID="c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.825647 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff\": container with ID starting with c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff not found: ID does not exist" containerID="c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.825666 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff"} err="failed to get container status \"c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff\": rpc error: code = NotFound desc = could not find container \"c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff\": container with ID starting with c366ae570a55e65baf8d6407c604b011367bbfa2fadafbf7ccbe31587ca221ff not found: ID does not exist" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.836200 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71c3c74-923f-4fe9-ad3b-b74e3ce55447-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.951908 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.979014 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.989542 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.989982 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="init" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990000 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="init" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.990019 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="sg-core" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990026 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="sg-core" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.990057 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="dnsmasq-dns" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990066 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="dnsmasq-dns" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.990076 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-central-agent" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990082 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-central-agent" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.990096 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-notification-agent" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990102 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-notification-agent" Dec 04 14:21:52 crc kubenswrapper[4715]: E1204 14:21:52.990111 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="proxy-httpd" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990117 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="proxy-httpd" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990333 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-notification-agent" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990352 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="sg-core" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990368 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="dnsmasq-dns" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990379 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="proxy-httpd" Dec 04 14:21:52 crc kubenswrapper[4715]: I1204 14:21:52.990388 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" containerName="ceilometer-central-agent" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.011189 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.011724 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.014445 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.014632 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.014669 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042306 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qqfh\" (UniqueName: \"kubernetes.io/projected/347af36d-eea9-4d43-96e3-c20d6915ff4d-kube-api-access-5qqfh\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/347af36d-eea9-4d43-96e3-c20d6915ff4d-run-httpd\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042457 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-config-data\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/347af36d-eea9-4d43-96e3-c20d6915ff4d-log-httpd\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042523 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042561 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-scripts\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.042625 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/347af36d-eea9-4d43-96e3-c20d6915ff4d-run-httpd\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143494 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-config-data\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143531 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/347af36d-eea9-4d43-96e3-c20d6915ff4d-log-httpd\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143581 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143618 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-scripts\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143675 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.143788 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qqfh\" (UniqueName: \"kubernetes.io/projected/347af36d-eea9-4d43-96e3-c20d6915ff4d-kube-api-access-5qqfh\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.144000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/347af36d-eea9-4d43-96e3-c20d6915ff4d-run-httpd\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.146346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/347af36d-eea9-4d43-96e3-c20d6915ff4d-log-httpd\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.148711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.149071 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-scripts\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.149343 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.149497 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.149566 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/347af36d-eea9-4d43-96e3-c20d6915ff4d-config-data\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.161556 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qqfh\" (UniqueName: \"kubernetes.io/projected/347af36d-eea9-4d43-96e3-c20d6915ff4d-kube-api-access-5qqfh\") pod \"ceilometer-0\" (UID: \"347af36d-eea9-4d43-96e3-c20d6915ff4d\") " pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.192963 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b71c3c74-923f-4fe9-ad3b-b74e3ce55447" path="/var/lib/kubelet/pods/b71c3c74-923f-4fe9-ad3b-b74e3ce55447/volumes" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.341422 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:21:53 crc kubenswrapper[4715]: I1204 14:21:53.792007 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:21:53 crc kubenswrapper[4715]: W1204 14:21:53.793454 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod347af36d_eea9_4d43_96e3_c20d6915ff4d.slice/crio-2c90a5b8ca7cbbcc19d9b38003ba02c4228b990034d7264e624ea1e703de1546 WatchSource:0}: Error finding container 2c90a5b8ca7cbbcc19d9b38003ba02c4228b990034d7264e624ea1e703de1546: Status 404 returned error can't find the container with id 2c90a5b8ca7cbbcc19d9b38003ba02c4228b990034d7264e624ea1e703de1546 Dec 04 14:21:54 crc kubenswrapper[4715]: I1204 14:21:54.444518 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-gvqf4" podUID="b2a21e37-a737-4818-a332-e2e8048314c3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: i/o timeout" Dec 04 14:21:54 crc kubenswrapper[4715]: I1204 14:21:54.636857 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"347af36d-eea9-4d43-96e3-c20d6915ff4d","Type":"ContainerStarted","Data":"2c90a5b8ca7cbbcc19d9b38003ba02c4228b990034d7264e624ea1e703de1546"} Dec 04 14:21:54 crc kubenswrapper[4715]: I1204 14:21:54.638593 4715 generic.go:334] "Generic (PLEG): container finished" podID="87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" containerID="e624bc6b5d84e244abb219db0871ef091414caa93421ce28f0ce179a22af2a24" exitCode=0 Dec 04 14:21:54 crc kubenswrapper[4715]: I1204 14:21:54.638625 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7hlzp" event={"ID":"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb","Type":"ContainerDied","Data":"e624bc6b5d84e244abb219db0871ef091414caa93421ce28f0ce179a22af2a24"} Dec 04 14:21:55 crc kubenswrapper[4715]: I1204 14:21:55.651221 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"347af36d-eea9-4d43-96e3-c20d6915ff4d","Type":"ContainerStarted","Data":"35d30be5f890415e546c9ffa4a6199ffbe678c6ce6f77b567c5dbdf8a0f1d99c"} Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.041672 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.193718 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.194918 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.216426 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-scripts\") pod \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.216747 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-config-data\") pod \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.216793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrbm7\" (UniqueName: \"kubernetes.io/projected/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-kube-api-access-rrbm7\") pod \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.216825 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-combined-ca-bundle\") pod \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\" (UID: \"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb\") " Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.221457 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-scripts" (OuterVolumeSpecName: "scripts") pod "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" (UID: "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.224385 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-kube-api-access-rrbm7" (OuterVolumeSpecName: "kube-api-access-rrbm7") pod "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" (UID: "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb"). InnerVolumeSpecName "kube-api-access-rrbm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.261367 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-config-data" (OuterVolumeSpecName: "config-data") pod "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" (UID: "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.275656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" (UID: "87e3b168-dd77-4ba1-bdd4-dd4072b30dfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.320122 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.320162 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.320172 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrbm7\" (UniqueName: \"kubernetes.io/projected/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-kube-api-access-rrbm7\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.320182 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.671649 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7hlzp" event={"ID":"87e3b168-dd77-4ba1-bdd4-dd4072b30dfb","Type":"ContainerDied","Data":"2f763d4f0c10fb4d5fcb877c5b73a5188cda55b751cc2ee78049b85cd5f05b47"} Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.671701 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f763d4f0c10fb4d5fcb877c5b73a5188cda55b751cc2ee78049b85cd5f05b47" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.671820 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7hlzp" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.678077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"347af36d-eea9-4d43-96e3-c20d6915ff4d","Type":"ContainerStarted","Data":"1fbad478f0d16fce82d90369d66980027dd2edda83c5c6c1c2136bd486a85ba1"} Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.678402 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"347af36d-eea9-4d43-96e3-c20d6915ff4d","Type":"ContainerStarted","Data":"f75dbba6c8a28d870f9a1d37e3b908905ef121151ac77430239f1478b38a1f59"} Dec 04 14:21:56 crc kubenswrapper[4715]: E1204 14:21:56.840408 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87e3b168_dd77_4ba1_bdd4_dd4072b30dfb.slice\": RecentStats: unable to find data in memory cache]" Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.874002 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.932598 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.932949 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1cfe0845-21fa-452e-be72-b6321946c018" containerName="nova-scheduler-scheduler" containerID="cri-o://fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" gracePeriod=30 Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.953268 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.953517 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-log" containerID="cri-o://a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9" gracePeriod=30 Dec 04 14:21:56 crc kubenswrapper[4715]: I1204 14:21:56.953659 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-metadata" containerID="cri-o://fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43" gracePeriod=30 Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.211441 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.211989 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.691430 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"347af36d-eea9-4d43-96e3-c20d6915ff4d","Type":"ContainerStarted","Data":"9618e138499f402d9ac050a80bc9a851aa44e996f8815a559dd7ac416d94296f"} Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.691765 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.694875 4715 generic.go:334] "Generic (PLEG): container finished" podID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerID="a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9" exitCode=143 Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.694954 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2","Type":"ContainerDied","Data":"a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9"} Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.695127 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-log" containerID="cri-o://b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4" gracePeriod=30 Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.695199 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-api" containerID="cri-o://90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771" gracePeriod=30 Dec 04 14:21:57 crc kubenswrapper[4715]: I1204 14:21:57.731283 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.264767173 podStartE2EDuration="5.73123656s" podCreationTimestamp="2025-12-04 14:21:52 +0000 UTC" firstStartedPulling="2025-12-04 14:21:53.796913002 +0000 UTC m=+1470.865631217" lastFinishedPulling="2025-12-04 14:21:57.263382389 +0000 UTC m=+1474.332100604" observedRunningTime="2025-12-04 14:21:57.719944904 +0000 UTC m=+1474.788663129" watchObservedRunningTime="2025-12-04 14:21:57.73123656 +0000 UTC m=+1474.799954785" Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.283575 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94 is running failed: container process not found" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.284509 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94 is running failed: container process not found" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.284970 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94 is running failed: container process not found" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.285068 4715 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1cfe0845-21fa-452e-be72-b6321946c018" containerName="nova-scheduler-scheduler" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.454602 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.575228 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-combined-ca-bundle\") pod \"1cfe0845-21fa-452e-be72-b6321946c018\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.575483 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-config-data\") pod \"1cfe0845-21fa-452e-be72-b6321946c018\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.575563 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x86sb\" (UniqueName: \"kubernetes.io/projected/1cfe0845-21fa-452e-be72-b6321946c018-kube-api-access-x86sb\") pod \"1cfe0845-21fa-452e-be72-b6321946c018\" (UID: \"1cfe0845-21fa-452e-be72-b6321946c018\") " Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.591267 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cfe0845-21fa-452e-be72-b6321946c018-kube-api-access-x86sb" (OuterVolumeSpecName: "kube-api-access-x86sb") pod "1cfe0845-21fa-452e-be72-b6321946c018" (UID: "1cfe0845-21fa-452e-be72-b6321946c018"). InnerVolumeSpecName "kube-api-access-x86sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.611950 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cfe0845-21fa-452e-be72-b6321946c018" (UID: "1cfe0845-21fa-452e-be72-b6321946c018"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.616295 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-config-data" (OuterVolumeSpecName: "config-data") pod "1cfe0845-21fa-452e-be72-b6321946c018" (UID: "1cfe0845-21fa-452e-be72-b6321946c018"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.678624 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.678654 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cfe0845-21fa-452e-be72-b6321946c018-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.678665 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x86sb\" (UniqueName: \"kubernetes.io/projected/1cfe0845-21fa-452e-be72-b6321946c018-kube-api-access-x86sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.706550 4715 generic.go:334] "Generic (PLEG): container finished" podID="1cfe0845-21fa-452e-be72-b6321946c018" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" exitCode=0 Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.706624 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1cfe0845-21fa-452e-be72-b6321946c018","Type":"ContainerDied","Data":"fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94"} Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.706657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1cfe0845-21fa-452e-be72-b6321946c018","Type":"ContainerDied","Data":"4e851d522a9d5f1d81b0e89a2fbca8198e71e9bf2b77fdf9d203460aeed613d3"} Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.706677 4715 scope.go:117] "RemoveContainer" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.706811 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.718636 4715 generic.go:334] "Generic (PLEG): container finished" podID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerID="b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4" exitCode=143 Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.718683 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a28e21c-f41b-4f6e-b2f5-303bc0828613","Type":"ContainerDied","Data":"b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4"} Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.748800 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.754422 4715 scope.go:117] "RemoveContainer" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.757291 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94\": container with ID starting with fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94 not found: ID does not exist" containerID="fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.757337 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94"} err="failed to get container status \"fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94\": rpc error: code = NotFound desc = could not find container \"fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94\": container with ID starting with fd7ab265bbd445456bee3d2d90ab11a8a70f9542672c19ee946a2933fb676b94 not found: ID does not exist" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.766129 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.779452 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.779917 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" containerName="nova-manage" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.779935 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" containerName="nova-manage" Dec 04 14:21:58 crc kubenswrapper[4715]: E1204 14:21:58.779980 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cfe0845-21fa-452e-be72-b6321946c018" containerName="nova-scheduler-scheduler" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.779987 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cfe0845-21fa-452e-be72-b6321946c018" containerName="nova-scheduler-scheduler" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.780190 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cfe0845-21fa-452e-be72-b6321946c018" containerName="nova-scheduler-scheduler" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.780221 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" containerName="nova-manage" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.780843 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.785603 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.797314 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.884227 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnh9c\" (UniqueName: \"kubernetes.io/projected/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-kube-api-access-lnh9c\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.884332 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-config-data\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.884415 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.987225 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnh9c\" (UniqueName: \"kubernetes.io/projected/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-kube-api-access-lnh9c\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.987299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-config-data\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.987371 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:58 crc kubenswrapper[4715]: I1204 14:21:58.992121 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-config-data\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:59 crc kubenswrapper[4715]: I1204 14:21:59.006451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnh9c\" (UniqueName: \"kubernetes.io/projected/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-kube-api-access-lnh9c\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:59 crc kubenswrapper[4715]: I1204 14:21:59.007215 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce\") " pod="openstack/nova-scheduler-0" Dec 04 14:21:59 crc kubenswrapper[4715]: I1204 14:21:59.103897 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:21:59 crc kubenswrapper[4715]: I1204 14:21:59.249140 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cfe0845-21fa-452e-be72-b6321946c018" path="/var/lib/kubelet/pods/1cfe0845-21fa-452e-be72-b6321946c018/volumes" Dec 04 14:21:59 crc kubenswrapper[4715]: I1204 14:21:59.629670 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:21:59 crc kubenswrapper[4715]: I1204 14:21:59.730760 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce","Type":"ContainerStarted","Data":"de344c12741b445c8a311ee7632e71760dde584276096e75628930cab6cfbe93"} Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.246072 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:33424->10.217.0.191:8775: read: connection reset by peer" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.246476 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:33426->10.217.0.191:8775: read: connection reset by peer" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.693634 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.753530 4715 generic.go:334] "Generic (PLEG): container finished" podID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerID="fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43" exitCode=0 Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.754025 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.754450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2","Type":"ContainerDied","Data":"fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43"} Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.755710 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2","Type":"ContainerDied","Data":"aa15f852cf93a01e088f32980b4cabb84d5fd6fa893b48194ddd254b7b3d1f73"} Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.755815 4715 scope.go:117] "RemoveContainer" containerID="fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.764091 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce","Type":"ContainerStarted","Data":"c41f41898db945b73ae4ef3b9689254714935dfceb57a59e8808c839b6c4e0de"} Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.792267 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.792243138 podStartE2EDuration="2.792243138s" podCreationTimestamp="2025-12-04 14:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:22:00.784614212 +0000 UTC m=+1477.853332427" watchObservedRunningTime="2025-12-04 14:22:00.792243138 +0000 UTC m=+1477.860961363" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.810351 4715 scope.go:117] "RemoveContainer" containerID="a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.837050 4715 scope.go:117] "RemoveContainer" containerID="fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43" Dec 04 14:22:00 crc kubenswrapper[4715]: E1204 14:22:00.837614 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43\": container with ID starting with fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43 not found: ID does not exist" containerID="fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.837734 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43"} err="failed to get container status \"fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43\": rpc error: code = NotFound desc = could not find container \"fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43\": container with ID starting with fffd07180f989845f74afb65ae25a0ba94860ad43e3afc59c416c6d41f02ad43 not found: ID does not exist" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.837853 4715 scope.go:117] "RemoveContainer" containerID="a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9" Dec 04 14:22:00 crc kubenswrapper[4715]: E1204 14:22:00.838989 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9\": container with ID starting with a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9 not found: ID does not exist" containerID="a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.839079 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9"} err="failed to get container status \"a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9\": rpc error: code = NotFound desc = could not find container \"a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9\": container with ID starting with a3ab7bc899356304a7b7c11efc66fddf589cfeaf5a9db48df6baa4373b97f3c9 not found: ID does not exist" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.851011 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-config-data\") pod \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.851117 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-logs\") pod \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.851153 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shnvs\" (UniqueName: \"kubernetes.io/projected/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-kube-api-access-shnvs\") pod \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.851181 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-nova-metadata-tls-certs\") pod \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.851256 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-combined-ca-bundle\") pod \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\" (UID: \"80383f63-e220-4ca3-9b8b-d7e2d8de9ab2\") " Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.853816 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-logs" (OuterVolumeSpecName: "logs") pod "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" (UID: "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.870441 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-kube-api-access-shnvs" (OuterVolumeSpecName: "kube-api-access-shnvs") pod "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" (UID: "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2"). InnerVolumeSpecName "kube-api-access-shnvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.889336 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" (UID: "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.904555 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-config-data" (OuterVolumeSpecName: "config-data") pod "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" (UID: "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.932714 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" (UID: "80383f63-e220-4ca3-9b8b-d7e2d8de9ab2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.954716 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.954756 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.954769 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shnvs\" (UniqueName: \"kubernetes.io/projected/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-kube-api-access-shnvs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.954783 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:00 crc kubenswrapper[4715]: I1204 14:22:00.954797 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.091889 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.107614 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.122415 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:22:01 crc kubenswrapper[4715]: E1204 14:22:01.122821 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-metadata" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.122840 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-metadata" Dec 04 14:22:01 crc kubenswrapper[4715]: E1204 14:22:01.122860 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-log" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.122867 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-log" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.123028 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-log" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.123055 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" containerName="nova-metadata-metadata" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.127893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.131289 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.135317 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.137014 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.193721 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80383f63-e220-4ca3-9b8b-d7e2d8de9ab2" path="/var/lib/kubelet/pods/80383f63-e220-4ca3-9b8b-d7e2d8de9ab2/volumes" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.260936 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwd2x\" (UniqueName: \"kubernetes.io/projected/62b5d096-e3b9-4fae-a697-d5a4b693514d-kube-api-access-xwd2x\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.261027 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5d096-e3b9-4fae-a697-d5a4b693514d-logs\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.261149 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.261173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-config-data\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.261226 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.362812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.362858 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-config-data\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.362917 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.362961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwd2x\" (UniqueName: \"kubernetes.io/projected/62b5d096-e3b9-4fae-a697-d5a4b693514d-kube-api-access-xwd2x\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.363016 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5d096-e3b9-4fae-a697-d5a4b693514d-logs\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.363406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5d096-e3b9-4fae-a697-d5a4b693514d-logs\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.367230 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.367709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-config-data\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.368098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5d096-e3b9-4fae-a697-d5a4b693514d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.385454 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwd2x\" (UniqueName: \"kubernetes.io/projected/62b5d096-e3b9-4fae-a697-d5a4b693514d-kube-api-access-xwd2x\") pod \"nova-metadata-0\" (UID: \"62b5d096-e3b9-4fae-a697-d5a4b693514d\") " pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.456125 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:22:01 crc kubenswrapper[4715]: I1204 14:22:01.934522 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:22:02 crc kubenswrapper[4715]: I1204 14:22:02.793932 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62b5d096-e3b9-4fae-a697-d5a4b693514d","Type":"ContainerStarted","Data":"ba7725460ae5708fe96749929c42e218200428dc842b9829f3b3f7581bf692fa"} Dec 04 14:22:02 crc kubenswrapper[4715]: I1204 14:22:02.794554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62b5d096-e3b9-4fae-a697-d5a4b693514d","Type":"ContainerStarted","Data":"d0e289faa9e4c00d6e93bc4e861ed5031daeb5e47f4a83417f3b1b436d07c764"} Dec 04 14:22:02 crc kubenswrapper[4715]: I1204 14:22:02.794571 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62b5d096-e3b9-4fae-a697-d5a4b693514d","Type":"ContainerStarted","Data":"7890e124d9b58d30b68635b2216f1ec0ca14a95a73e307956528b9e786e3576d"} Dec 04 14:22:02 crc kubenswrapper[4715]: I1204 14:22:02.817834 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.817811265 podStartE2EDuration="1.817811265s" podCreationTimestamp="2025-12-04 14:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:22:02.81653493 +0000 UTC m=+1479.885253165" watchObservedRunningTime="2025-12-04 14:22:02.817811265 +0000 UTC m=+1479.886529480" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.105089 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.651338 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.743996 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-combined-ca-bundle\") pod \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.744174 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-config-data\") pod \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.744276 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlpst\" (UniqueName: \"kubernetes.io/projected/7a28e21c-f41b-4f6e-b2f5-303bc0828613-kube-api-access-dlpst\") pod \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.744307 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-public-tls-certs\") pod \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.744366 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-internal-tls-certs\") pod \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.744424 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a28e21c-f41b-4f6e-b2f5-303bc0828613-logs\") pod \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\" (UID: \"7a28e21c-f41b-4f6e-b2f5-303bc0828613\") " Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.745291 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a28e21c-f41b-4f6e-b2f5-303bc0828613-logs" (OuterVolumeSpecName: "logs") pod "7a28e21c-f41b-4f6e-b2f5-303bc0828613" (UID: "7a28e21c-f41b-4f6e-b2f5-303bc0828613"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.746078 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a28e21c-f41b-4f6e-b2f5-303bc0828613-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.754590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a28e21c-f41b-4f6e-b2f5-303bc0828613-kube-api-access-dlpst" (OuterVolumeSpecName: "kube-api-access-dlpst") pod "7a28e21c-f41b-4f6e-b2f5-303bc0828613" (UID: "7a28e21c-f41b-4f6e-b2f5-303bc0828613"). InnerVolumeSpecName "kube-api-access-dlpst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.782678 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a28e21c-f41b-4f6e-b2f5-303bc0828613" (UID: "7a28e21c-f41b-4f6e-b2f5-303bc0828613"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.789096 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-config-data" (OuterVolumeSpecName: "config-data") pod "7a28e21c-f41b-4f6e-b2f5-303bc0828613" (UID: "7a28e21c-f41b-4f6e-b2f5-303bc0828613"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.813755 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7a28e21c-f41b-4f6e-b2f5-303bc0828613" (UID: "7a28e21c-f41b-4f6e-b2f5-303bc0828613"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.816108 4715 generic.go:334] "Generic (PLEG): container finished" podID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerID="90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771" exitCode=0 Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.816161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a28e21c-f41b-4f6e-b2f5-303bc0828613","Type":"ContainerDied","Data":"90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771"} Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.816251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a28e21c-f41b-4f6e-b2f5-303bc0828613","Type":"ContainerDied","Data":"328b95b80dc8687b515c93c4eb3bb34f83d25d6abbaeddaf37c411e954463c3c"} Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.816283 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.816285 4715 scope.go:117] "RemoveContainer" containerID="90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.821228 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7a28e21c-f41b-4f6e-b2f5-303bc0828613" (UID: "7a28e21c-f41b-4f6e-b2f5-303bc0828613"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.848213 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.848254 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlpst\" (UniqueName: \"kubernetes.io/projected/7a28e21c-f41b-4f6e-b2f5-303bc0828613-kube-api-access-dlpst\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.848268 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.848279 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.848289 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a28e21c-f41b-4f6e-b2f5-303bc0828613-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.891229 4715 scope.go:117] "RemoveContainer" containerID="b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.918027 4715 scope.go:117] "RemoveContainer" containerID="90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771" Dec 04 14:22:04 crc kubenswrapper[4715]: E1204 14:22:04.918493 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771\": container with ID starting with 90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771 not found: ID does not exist" containerID="90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.918559 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771"} err="failed to get container status \"90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771\": rpc error: code = NotFound desc = could not find container \"90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771\": container with ID starting with 90c6074f2aaf250b2369876ee3badd1cb8a8648c66ff206a242144b1c6aee771 not found: ID does not exist" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.918633 4715 scope.go:117] "RemoveContainer" containerID="b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4" Dec 04 14:22:04 crc kubenswrapper[4715]: E1204 14:22:04.919083 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4\": container with ID starting with b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4 not found: ID does not exist" containerID="b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4" Dec 04 14:22:04 crc kubenswrapper[4715]: I1204 14:22:04.919106 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4"} err="failed to get container status \"b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4\": rpc error: code = NotFound desc = could not find container \"b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4\": container with ID starting with b5223ebc69bd5f3f192cabd33a89e2cd47fa5e6221efbdc0c3194acde03490c4 not found: ID does not exist" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.153087 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.166216 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.176369 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:22:05 crc kubenswrapper[4715]: E1204 14:22:05.176816 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-log" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.176833 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-log" Dec 04 14:22:05 crc kubenswrapper[4715]: E1204 14:22:05.176856 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-api" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.176862 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-api" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.177163 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-api" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.177177 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" containerName="nova-api-log" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.179669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.186375 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.186809 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.186980 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.211222 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a28e21c-f41b-4f6e-b2f5-303bc0828613" path="/var/lib/kubelet/pods/7a28e21c-f41b-4f6e-b2f5-303bc0828613/volumes" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.211955 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.256372 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxdrp\" (UniqueName: \"kubernetes.io/projected/9ec54636-c541-4292-b5b9-7a6215b8828e-kube-api-access-cxdrp\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.256550 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.256635 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ec54636-c541-4292-b5b9-7a6215b8828e-logs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.256675 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-config-data\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.256737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.256776 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.358179 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.358254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ec54636-c541-4292-b5b9-7a6215b8828e-logs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.358285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-config-data\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.358338 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.358384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.358475 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxdrp\" (UniqueName: \"kubernetes.io/projected/9ec54636-c541-4292-b5b9-7a6215b8828e-kube-api-access-cxdrp\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.359282 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ec54636-c541-4292-b5b9-7a6215b8828e-logs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.363313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.363665 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.365612 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.366520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ec54636-c541-4292-b5b9-7a6215b8828e-config-data\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.378942 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxdrp\" (UniqueName: \"kubernetes.io/projected/9ec54636-c541-4292-b5b9-7a6215b8828e-kube-api-access-cxdrp\") pod \"nova-api-0\" (UID: \"9ec54636-c541-4292-b5b9-7a6215b8828e\") " pod="openstack/nova-api-0" Dec 04 14:22:05 crc kubenswrapper[4715]: I1204 14:22:05.516767 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.008692 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:22:06 crc kubenswrapper[4715]: W1204 14:22:06.018399 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ec54636_c541_4292_b5b9_7a6215b8828e.slice/crio-7163bcb4fd8007122e76255833213a329a8600b1a9757ebb49f717f8b59ae6e3 WatchSource:0}: Error finding container 7163bcb4fd8007122e76255833213a329a8600b1a9757ebb49f717f8b59ae6e3: Status 404 returned error can't find the container with id 7163bcb4fd8007122e76255833213a329a8600b1a9757ebb49f717f8b59ae6e3 Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.457274 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.457658 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.841267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9ec54636-c541-4292-b5b9-7a6215b8828e","Type":"ContainerStarted","Data":"f6efaad68a37a64f25163f10e1ebdda72a8e936ba57545858311bd94431917f7"} Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.841325 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9ec54636-c541-4292-b5b9-7a6215b8828e","Type":"ContainerStarted","Data":"127e08175ed295c49eaf4a616d39443d597f423b96074c776a700046f10e76ee"} Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.841338 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9ec54636-c541-4292-b5b9-7a6215b8828e","Type":"ContainerStarted","Data":"7163bcb4fd8007122e76255833213a329a8600b1a9757ebb49f717f8b59ae6e3"} Dec 04 14:22:06 crc kubenswrapper[4715]: I1204 14:22:06.869931 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.869905502 podStartE2EDuration="1.869905502s" podCreationTimestamp="2025-12-04 14:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:22:06.861715503 +0000 UTC m=+1483.930433738" watchObservedRunningTime="2025-12-04 14:22:06.869905502 +0000 UTC m=+1483.938623717" Dec 04 14:22:08 crc kubenswrapper[4715]: I1204 14:22:08.758655 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:22:08 crc kubenswrapper[4715]: I1204 14:22:08.758972 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:22:08 crc kubenswrapper[4715]: I1204 14:22:08.759021 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:22:08 crc kubenswrapper[4715]: I1204 14:22:08.759758 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f8cf4969e2a565cf9483d0133bd3e156b294d54b874a31409400fd261c14615"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:22:08 crc kubenswrapper[4715]: I1204 14:22:08.759818 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://5f8cf4969e2a565cf9483d0133bd3e156b294d54b874a31409400fd261c14615" gracePeriod=600 Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.104740 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.148006 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.869324 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="5f8cf4969e2a565cf9483d0133bd3e156b294d54b874a31409400fd261c14615" exitCode=0 Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.869397 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"5f8cf4969e2a565cf9483d0133bd3e156b294d54b874a31409400fd261c14615"} Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.870658 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b"} Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.870678 4715 scope.go:117] "RemoveContainer" containerID="d601d4f16680a55d11b4648fd80ee11fd0f65f95595398da0f63f96f86a77a1f" Dec 04 14:22:09 crc kubenswrapper[4715]: I1204 14:22:09.906299 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:22:11 crc kubenswrapper[4715]: I1204 14:22:11.457708 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:22:11 crc kubenswrapper[4715]: I1204 14:22:11.458298 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:22:12 crc kubenswrapper[4715]: I1204 14:22:12.473553 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="62b5d096-e3b9-4fae-a697-d5a4b693514d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:22:12 crc kubenswrapper[4715]: I1204 14:22:12.473534 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="62b5d096-e3b9-4fae-a697-d5a4b693514d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:22:15 crc kubenswrapper[4715]: I1204 14:22:15.518002 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:22:15 crc kubenswrapper[4715]: I1204 14:22:15.518524 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:22:16 crc kubenswrapper[4715]: I1204 14:22:16.532436 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9ec54636-c541-4292-b5b9-7a6215b8828e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:22:16 crc kubenswrapper[4715]: I1204 14:22:16.532437 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9ec54636-c541-4292-b5b9-7a6215b8828e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:22:21 crc kubenswrapper[4715]: I1204 14:22:21.464411 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:22:21 crc kubenswrapper[4715]: I1204 14:22:21.465103 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:22:21 crc kubenswrapper[4715]: I1204 14:22:21.470969 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:22:21 crc kubenswrapper[4715]: I1204 14:22:21.472768 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:22:23 crc kubenswrapper[4715]: I1204 14:22:23.364884 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:22:25 crc kubenswrapper[4715]: I1204 14:22:25.523803 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:22:25 crc kubenswrapper[4715]: I1204 14:22:25.524471 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:22:25 crc kubenswrapper[4715]: I1204 14:22:25.524817 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:22:25 crc kubenswrapper[4715]: I1204 14:22:25.524836 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:22:25 crc kubenswrapper[4715]: I1204 14:22:25.532243 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:22:25 crc kubenswrapper[4715]: I1204 14:22:25.533095 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:22:34 crc kubenswrapper[4715]: I1204 14:22:34.036257 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:22:34 crc kubenswrapper[4715]: I1204 14:22:34.934582 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:22:37 crc kubenswrapper[4715]: I1204 14:22:37.732512 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6l5jl"] Dec 04 14:22:37 crc kubenswrapper[4715]: I1204 14:22:37.735632 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:37 crc kubenswrapper[4715]: I1204 14:22:37.750844 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6l5jl"] Dec 04 14:22:37 crc kubenswrapper[4715]: I1204 14:22:37.910644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlrb5\" (UniqueName: \"kubernetes.io/projected/177675e9-6c99-4f27-9545-d223714124ef-kube-api-access-jlrb5\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:37 crc kubenswrapper[4715]: I1204 14:22:37.910715 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-utilities\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:37 crc kubenswrapper[4715]: I1204 14:22:37.910797 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-catalog-content\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.012817 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlrb5\" (UniqueName: \"kubernetes.io/projected/177675e9-6c99-4f27-9545-d223714124ef-kube-api-access-jlrb5\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.012901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-utilities\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.013013 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-catalog-content\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.013479 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-utilities\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.013520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-catalog-content\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.042151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlrb5\" (UniqueName: \"kubernetes.io/projected/177675e9-6c99-4f27-9545-d223714124ef-kube-api-access-jlrb5\") pod \"redhat-operators-6l5jl\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.070716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.562125 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6l5jl"] Dec 04 14:22:38 crc kubenswrapper[4715]: I1204 14:22:38.836001 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="rabbitmq" containerID="cri-o://8ab6c67abc9e697ad6cc5592319e9baccda512a795b344eb06ca040df3be5cd9" gracePeriod=604796 Dec 04 14:22:39 crc kubenswrapper[4715]: I1204 14:22:39.141284 4715 generic.go:334] "Generic (PLEG): container finished" podID="177675e9-6c99-4f27-9545-d223714124ef" containerID="9d7201d42e6c7ebecc662312fcd21afebe9ba11a27954af1990f8a0f4e1c1cc4" exitCode=0 Dec 04 14:22:39 crc kubenswrapper[4715]: I1204 14:22:39.141498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerDied","Data":"9d7201d42e6c7ebecc662312fcd21afebe9ba11a27954af1990f8a0f4e1c1cc4"} Dec 04 14:22:39 crc kubenswrapper[4715]: I1204 14:22:39.141562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerStarted","Data":"822ea81cac84443106e2ec2fde9874ef37e78d08913e44620755499a37e9667e"} Dec 04 14:22:39 crc kubenswrapper[4715]: I1204 14:22:39.377543 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 04 14:22:40 crc kubenswrapper[4715]: I1204 14:22:40.085275 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="rabbitmq" containerID="cri-o://6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71" gracePeriod=604795 Dec 04 14:22:40 crc kubenswrapper[4715]: I1204 14:22:40.153698 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerStarted","Data":"affb94972edcec86554f1d0a4f8b4a6df36fec99e74ec144de2d7b08e12923e4"} Dec 04 14:22:41 crc kubenswrapper[4715]: I1204 14:22:41.163576 4715 generic.go:334] "Generic (PLEG): container finished" podID="177675e9-6c99-4f27-9545-d223714124ef" containerID="affb94972edcec86554f1d0a4f8b4a6df36fec99e74ec144de2d7b08e12923e4" exitCode=0 Dec 04 14:22:41 crc kubenswrapper[4715]: I1204 14:22:41.163691 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerDied","Data":"affb94972edcec86554f1d0a4f8b4a6df36fec99e74ec144de2d7b08e12923e4"} Dec 04 14:22:42 crc kubenswrapper[4715]: I1204 14:22:42.176088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerStarted","Data":"7aa048a7ad311ca45e7c30bb9e10f0cf82facbbf2e48c3b5f2c97c67f19bde64"} Dec 04 14:22:42 crc kubenswrapper[4715]: I1204 14:22:42.197597 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6l5jl" podStartSLOduration=2.506988366 podStartE2EDuration="5.19757997s" podCreationTimestamp="2025-12-04 14:22:37 +0000 UTC" firstStartedPulling="2025-12-04 14:22:39.143394768 +0000 UTC m=+1516.212112983" lastFinishedPulling="2025-12-04 14:22:41.833986372 +0000 UTC m=+1518.902704587" observedRunningTime="2025-12-04 14:22:42.192954816 +0000 UTC m=+1519.261673031" watchObservedRunningTime="2025-12-04 14:22:42.19757997 +0000 UTC m=+1519.266298185" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.220951 4715 generic.go:334] "Generic (PLEG): container finished" podID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerID="8ab6c67abc9e697ad6cc5592319e9baccda512a795b344eb06ca040df3be5cd9" exitCode=0 Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.221069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"440f7f8d-a89d-4d6d-8280-c56fede1c91c","Type":"ContainerDied","Data":"8ab6c67abc9e697ad6cc5592319e9baccda512a795b344eb06ca040df3be5cd9"} Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.805910 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.982780 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-tls\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.982832 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-plugins\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.982892 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/440f7f8d-a89d-4d6d-8280-c56fede1c91c-pod-info\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.982929 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.982950 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-server-conf\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983005 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4rrs\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-kube-api-access-m4rrs\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983093 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-config-data\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983178 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/440f7f8d-a89d-4d6d-8280-c56fede1c91c-erlang-cookie-secret\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-plugins-conf\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983630 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983849 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-erlang-cookie\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.983882 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-confd\") pod \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\" (UID: \"440f7f8d-a89d-4d6d-8280-c56fede1c91c\") " Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.984119 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.984235 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.984789 4715 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.984810 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.984824 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.991178 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/440f7f8d-a89d-4d6d-8280-c56fede1c91c-pod-info" (OuterVolumeSpecName: "pod-info") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.991221 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/440f7f8d-a89d-4d6d-8280-c56fede1c91c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.991276 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.991291 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:46 crc kubenswrapper[4715]: I1204 14:22:46.991827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-kube-api-access-m4rrs" (OuterVolumeSpecName: "kube-api-access-m4rrs") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "kube-api-access-m4rrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.026688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-config-data" (OuterVolumeSpecName: "config-data") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.072250 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-server-conf" (OuterVolumeSpecName: "server-conf") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.086926 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4rrs\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-kube-api-access-m4rrs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.086982 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.086997 4715 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/440f7f8d-a89d-4d6d-8280-c56fede1c91c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.087005 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.087015 4715 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/440f7f8d-a89d-4d6d-8280-c56fede1c91c-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.087062 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.087075 4715 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/440f7f8d-a89d-4d6d-8280-c56fede1c91c-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.128679 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.129518 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "440f7f8d-a89d-4d6d-8280-c56fede1c91c" (UID: "440f7f8d-a89d-4d6d-8280-c56fede1c91c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.190941 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/440f7f8d-a89d-4d6d-8280-c56fede1c91c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.190990 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.240598 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"440f7f8d-a89d-4d6d-8280-c56fede1c91c","Type":"ContainerDied","Data":"c7ba1830e98359f3caa67efc3663c83e3ad2d8306c0af40e367d1ab878e8f775"} Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.240658 4715 scope.go:117] "RemoveContainer" containerID="8ab6c67abc9e697ad6cc5592319e9baccda512a795b344eb06ca040df3be5cd9" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.240671 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.276456 4715 scope.go:117] "RemoveContainer" containerID="70e712e943ad0c774cffe640698edc91acb2e0c08db4632f0be92d036d4b9251" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.293531 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.316911 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.342711 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:22:47 crc kubenswrapper[4715]: E1204 14:22:47.343191 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="rabbitmq" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.343209 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="rabbitmq" Dec 04 14:22:47 crc kubenswrapper[4715]: E1204 14:22:47.343237 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="setup-container" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.343243 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="setup-container" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.343439 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" containerName="rabbitmq" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.344670 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.348391 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.348533 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.348757 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.348809 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.349784 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tcnqh" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.349885 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.351954 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.354583 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499463 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499562 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499591 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-config-data\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499653 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499703 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499887 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.499956 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.500099 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkzgv\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-kube-api-access-hkzgv\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.524499 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-l4m5p"] Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.526603 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.528638 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.537668 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-l4m5p"] Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602330 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602534 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psnxx\" (UniqueName: \"kubernetes.io/projected/d7b82945-5f57-43bb-bb84-6d112339d87f-kube-api-access-psnxx\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-config-data\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602674 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602695 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602785 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602837 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-config\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602888 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602912 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.602966 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.603007 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.603329 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.603730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.603949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.605246 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.603072 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkzgv\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-kube-api-access-hkzgv\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.605878 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.606730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.607003 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-config-data\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.608806 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.609862 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.615015 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.622520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkzgv\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-kube-api-access-hkzgv\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.632110 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72cfc89d-7970-45e9-be5f-3f5f5d3f520d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.654313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"72cfc89d-7970-45e9-be5f-3f5f5d3f520d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.668096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707610 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707721 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-config\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707741 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707785 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707850 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707878 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psnxx\" (UniqueName: \"kubernetes.io/projected/d7b82945-5f57-43bb-bb84-6d112339d87f-kube-api-access-psnxx\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.707910 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.709085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.709163 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.709623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-config\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.710252 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.710389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.710966 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.744723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psnxx\" (UniqueName: \"kubernetes.io/projected/d7b82945-5f57-43bb-bb84-6d112339d87f-kube-api-access-psnxx\") pod \"dnsmasq-dns-79bd4cc8c9-l4m5p\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:47 crc kubenswrapper[4715]: I1204 14:22:47.885965 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.003618 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.070951 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.071413 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.114955 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-tls\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115132 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-plugins-conf\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115196 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-config-data\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-server-conf\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115264 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-pod-info\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115289 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrt59\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-kube-api-access-zrt59\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-plugins\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115339 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-confd\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115378 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-erlang-cookie-secret\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115398 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-erlang-cookie\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.115423 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\" (UID: \"d34885f9-cf98-4b1a-bbf3-c7af62f5f273\") " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.121248 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.122289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-kube-api-access-zrt59" (OuterVolumeSpecName: "kube-api-access-zrt59") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "kube-api-access-zrt59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.124203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.124442 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.125720 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.126722 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.126937 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-pod-info" (OuterVolumeSpecName: "pod-info") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.144078 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.150017 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.158985 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-config-data" (OuterVolumeSpecName: "config-data") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218105 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrt59\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-kube-api-access-zrt59\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218135 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218152 4715 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218164 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218199 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218212 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218224 4715 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218235 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.218245 4715 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.230018 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-server-conf" (OuterVolumeSpecName: "server-conf") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.241505 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.253247 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.256461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"72cfc89d-7970-45e9-be5f-3f5f5d3f520d","Type":"ContainerStarted","Data":"8534705d02462f8425d1eccb71aade701cbe8d0894244a91f83bb536840bcc14"} Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.257967 4715 generic.go:334] "Generic (PLEG): container finished" podID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerID="6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71" exitCode=0 Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.259135 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.259769 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34885f9-cf98-4b1a-bbf3-c7af62f5f273","Type":"ContainerDied","Data":"6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71"} Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.259796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34885f9-cf98-4b1a-bbf3-c7af62f5f273","Type":"ContainerDied","Data":"6563d11eed8877e6b691cd75b8a8b74020e65a4f46f79710abdbb14d6133e313"} Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.259812 4715 scope.go:117] "RemoveContainer" containerID="6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.284420 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d34885f9-cf98-4b1a-bbf3-c7af62f5f273" (UID: "d34885f9-cf98-4b1a-bbf3-c7af62f5f273"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.286640 4715 scope.go:117] "RemoveContainer" containerID="41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.312275 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.319666 4715 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.319704 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34885f9-cf98-4b1a-bbf3-c7af62f5f273-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.319717 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.345220 4715 scope.go:117] "RemoveContainer" containerID="6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71" Dec 04 14:22:48 crc kubenswrapper[4715]: E1204 14:22:48.345851 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71\": container with ID starting with 6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71 not found: ID does not exist" containerID="6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.345974 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71"} err="failed to get container status \"6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71\": rpc error: code = NotFound desc = could not find container \"6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71\": container with ID starting with 6ff3cc47355d6935908cc0f75b446e130fcbd8b961dbb04a094ec175ceca9b71 not found: ID does not exist" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.346007 4715 scope.go:117] "RemoveContainer" containerID="41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e" Dec 04 14:22:48 crc kubenswrapper[4715]: E1204 14:22:48.346555 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e\": container with ID starting with 41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e not found: ID does not exist" containerID="41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.346747 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e"} err="failed to get container status \"41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e\": rpc error: code = NotFound desc = could not find container \"41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e\": container with ID starting with 41f66c3b0fef1e5bc8e2c2692abd37af609884ff288905b287fb64885601557e not found: ID does not exist" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.393090 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6l5jl"] Dec 04 14:22:48 crc kubenswrapper[4715]: W1204 14:22:48.442813 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7b82945_5f57_43bb_bb84_6d112339d87f.slice/crio-55b15e4cc4588ed49e2bcdd26ad89899467312c3bcd89c88463481b0eb8f19a0 WatchSource:0}: Error finding container 55b15e4cc4588ed49e2bcdd26ad89899467312c3bcd89c88463481b0eb8f19a0: Status 404 returned error can't find the container with id 55b15e4cc4588ed49e2bcdd26ad89899467312c3bcd89c88463481b0eb8f19a0 Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.449505 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-l4m5p"] Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.619662 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.629870 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.657914 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:22:48 crc kubenswrapper[4715]: E1204 14:22:48.658371 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="rabbitmq" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.658390 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="rabbitmq" Dec 04 14:22:48 crc kubenswrapper[4715]: E1204 14:22:48.658425 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="setup-container" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.658432 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="setup-container" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.658624 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" containerName="rabbitmq" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.659875 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.661715 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.662708 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.663068 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.663240 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.663392 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-s8rcz" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.663539 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.665654 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.674490 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830621 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830720 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830740 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830833 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830920 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830936 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.830989 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h5dw\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-kube-api-access-5h5dw\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.831019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932282 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932386 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932409 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932486 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932518 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932559 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932576 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.932619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h5dw\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-kube-api-access-5h5dw\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.933339 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.933733 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.933779 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.934190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.934238 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.934729 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.938882 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.938988 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.940392 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.940565 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.952851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h5dw\" (UniqueName: \"kubernetes.io/projected/45d230ae-3ae5-4d87-8979-bbd1fd8651cd-kube-api-access-5h5dw\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:48 crc kubenswrapper[4715]: I1204 14:22:48.972956 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45d230ae-3ae5-4d87-8979-bbd1fd8651cd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.158768 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.205995 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="440f7f8d-a89d-4d6d-8280-c56fede1c91c" path="/var/lib/kubelet/pods/440f7f8d-a89d-4d6d-8280-c56fede1c91c/volumes" Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.207204 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d34885f9-cf98-4b1a-bbf3-c7af62f5f273" path="/var/lib/kubelet/pods/d34885f9-cf98-4b1a-bbf3-c7af62f5f273/volumes" Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.282007 4715 generic.go:334] "Generic (PLEG): container finished" podID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerID="46b21ea32285b0757d68c0f98b53b4b7083d2c108158bb922bc283eb1a7910d0" exitCode=0 Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.282104 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" event={"ID":"d7b82945-5f57-43bb-bb84-6d112339d87f","Type":"ContainerDied","Data":"46b21ea32285b0757d68c0f98b53b4b7083d2c108158bb922bc283eb1a7910d0"} Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.282230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" event={"ID":"d7b82945-5f57-43bb-bb84-6d112339d87f","Type":"ContainerStarted","Data":"55b15e4cc4588ed49e2bcdd26ad89899467312c3bcd89c88463481b0eb8f19a0"} Dec 04 14:22:49 crc kubenswrapper[4715]: I1204 14:22:49.741275 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:22:50 crc kubenswrapper[4715]: I1204 14:22:50.297595 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45d230ae-3ae5-4d87-8979-bbd1fd8651cd","Type":"ContainerStarted","Data":"56c538d17e3129c3333e3f1377c1563072aa07b6787d0d4b1784ebf67eb4a115"} Dec 04 14:22:50 crc kubenswrapper[4715]: I1204 14:22:50.300706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"72cfc89d-7970-45e9-be5f-3f5f5d3f520d","Type":"ContainerStarted","Data":"db4e68667a1ccefd7cac3522c126dcbb29b5c97e9397f0a348af19c78a408075"} Dec 04 14:22:50 crc kubenswrapper[4715]: I1204 14:22:50.303070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" event={"ID":"d7b82945-5f57-43bb-bb84-6d112339d87f","Type":"ContainerStarted","Data":"173c01970ca25a60a0b38a2bdbb0a44ea195aaea53077bc68c8a8de757b3b550"} Dec 04 14:22:50 crc kubenswrapper[4715]: I1204 14:22:50.303219 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6l5jl" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="registry-server" containerID="cri-o://7aa048a7ad311ca45e7c30bb9e10f0cf82facbbf2e48c3b5f2c97c67f19bde64" gracePeriod=2 Dec 04 14:22:50 crc kubenswrapper[4715]: I1204 14:22:50.360337 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" podStartSLOduration=3.360318062 podStartE2EDuration="3.360318062s" podCreationTimestamp="2025-12-04 14:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:22:50.355695268 +0000 UTC m=+1527.424413503" watchObservedRunningTime="2025-12-04 14:22:50.360318062 +0000 UTC m=+1527.429036277" Dec 04 14:22:51 crc kubenswrapper[4715]: I1204 14:22:51.311660 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.325291 4715 generic.go:334] "Generic (PLEG): container finished" podID="177675e9-6c99-4f27-9545-d223714124ef" containerID="7aa048a7ad311ca45e7c30bb9e10f0cf82facbbf2e48c3b5f2c97c67f19bde64" exitCode=0 Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.325422 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerDied","Data":"7aa048a7ad311ca45e7c30bb9e10f0cf82facbbf2e48c3b5f2c97c67f19bde64"} Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.329844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45d230ae-3ae5-4d87-8979-bbd1fd8651cd","Type":"ContainerStarted","Data":"f9e93287d1381b1cee1ec7dc34dac764b31d9bf7e7f2d0ff0d4517f97eefce2a"} Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.724012 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.822811 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-utilities\") pod \"177675e9-6c99-4f27-9545-d223714124ef\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.823185 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-catalog-content\") pod \"177675e9-6c99-4f27-9545-d223714124ef\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.823356 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlrb5\" (UniqueName: \"kubernetes.io/projected/177675e9-6c99-4f27-9545-d223714124ef-kube-api-access-jlrb5\") pod \"177675e9-6c99-4f27-9545-d223714124ef\" (UID: \"177675e9-6c99-4f27-9545-d223714124ef\") " Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.824086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-utilities" (OuterVolumeSpecName: "utilities") pod "177675e9-6c99-4f27-9545-d223714124ef" (UID: "177675e9-6c99-4f27-9545-d223714124ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.829898 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/177675e9-6c99-4f27-9545-d223714124ef-kube-api-access-jlrb5" (OuterVolumeSpecName: "kube-api-access-jlrb5") pod "177675e9-6c99-4f27-9545-d223714124ef" (UID: "177675e9-6c99-4f27-9545-d223714124ef"). InnerVolumeSpecName "kube-api-access-jlrb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.926199 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.926404 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlrb5\" (UniqueName: \"kubernetes.io/projected/177675e9-6c99-4f27-9545-d223714124ef-kube-api-access-jlrb5\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:52 crc kubenswrapper[4715]: I1204 14:22:52.932415 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "177675e9-6c99-4f27-9545-d223714124ef" (UID: "177675e9-6c99-4f27-9545-d223714124ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.028403 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/177675e9-6c99-4f27-9545-d223714124ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.341566 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l5jl" Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.341559 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l5jl" event={"ID":"177675e9-6c99-4f27-9545-d223714124ef","Type":"ContainerDied","Data":"822ea81cac84443106e2ec2fde9874ef37e78d08913e44620755499a37e9667e"} Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.341644 4715 scope.go:117] "RemoveContainer" containerID="7aa048a7ad311ca45e7c30bb9e10f0cf82facbbf2e48c3b5f2c97c67f19bde64" Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.368803 4715 scope.go:117] "RemoveContainer" containerID="affb94972edcec86554f1d0a4f8b4a6df36fec99e74ec144de2d7b08e12923e4" Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.372170 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6l5jl"] Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.381402 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6l5jl"] Dec 04 14:22:53 crc kubenswrapper[4715]: I1204 14:22:53.393819 4715 scope.go:117] "RemoveContainer" containerID="9d7201d42e6c7ebecc662312fcd21afebe9ba11a27954af1990f8a0f4e1c1cc4" Dec 04 14:22:55 crc kubenswrapper[4715]: I1204 14:22:55.193874 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="177675e9-6c99-4f27-9545-d223714124ef" path="/var/lib/kubelet/pods/177675e9-6c99-4f27-9545-d223714124ef/volumes" Dec 04 14:22:57 crc kubenswrapper[4715]: I1204 14:22:57.887865 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:22:57 crc kubenswrapper[4715]: I1204 14:22:57.981533 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4wwbd"] Dec 04 14:22:57 crc kubenswrapper[4715]: I1204 14:22:57.981820 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerName="dnsmasq-dns" containerID="cri-o://5c2ecfd532559c3fe36e2edcdb4dc3f60982356d9bd62a904f2e04d27f78619f" gracePeriod=10 Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.297605 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-rbrt6"] Dec 04 14:22:58 crc kubenswrapper[4715]: E1204 14:22:58.298670 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="extract-content" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.298689 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="extract-content" Dec 04 14:22:58 crc kubenswrapper[4715]: E1204 14:22:58.298717 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="registry-server" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.298723 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="registry-server" Dec 04 14:22:58 crc kubenswrapper[4715]: E1204 14:22:58.298732 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="extract-utilities" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.298738 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="extract-utilities" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.298933 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="177675e9-6c99-4f27-9545-d223714124ef" containerName="registry-server" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.300097 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333576 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333612 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gbt8\" (UniqueName: \"kubernetes.io/projected/dfe175eb-a594-4a23-a057-b3cbc53cad3e-kube-api-access-9gbt8\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333638 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333672 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-dns-svc\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333739 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-config\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.333758 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.338977 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-rbrt6"] Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.400377 4715 generic.go:334] "Generic (PLEG): container finished" podID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerID="5c2ecfd532559c3fe36e2edcdb4dc3f60982356d9bd62a904f2e04d27f78619f" exitCode=0 Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.400432 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" event={"ID":"8e7ccc35-38ad-4677-9cf3-305bc0d322c4","Type":"ContainerDied","Data":"5c2ecfd532559c3fe36e2edcdb4dc3f60982356d9bd62a904f2e04d27f78619f"} Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.434957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-dns-svc\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.435100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-config\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.435133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.435220 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.435262 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.435299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gbt8\" (UniqueName: \"kubernetes.io/projected/dfe175eb-a594-4a23-a057-b3cbc53cad3e-kube-api-access-9gbt8\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.435328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.436104 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-dns-svc\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.436635 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.436965 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-config\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.437755 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.437799 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.437846 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe175eb-a594-4a23-a057-b3cbc53cad3e-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.480020 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gbt8\" (UniqueName: \"kubernetes.io/projected/dfe175eb-a594-4a23-a057-b3cbc53cad3e-kube-api-access-9gbt8\") pod \"dnsmasq-dns-55478c4467-rbrt6\" (UID: \"dfe175eb-a594-4a23-a057-b3cbc53cad3e\") " pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.622693 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.761704 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.851706 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-svc\") pod \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.851807 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-swift-storage-0\") pod \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.851855 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-config\") pod \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.851902 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5dfs\" (UniqueName: \"kubernetes.io/projected/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-kube-api-access-q5dfs\") pod \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.852024 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-nb\") pod \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.853106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-sb\") pod \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\" (UID: \"8e7ccc35-38ad-4677-9cf3-305bc0d322c4\") " Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.879874 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-kube-api-access-q5dfs" (OuterVolumeSpecName: "kube-api-access-q5dfs") pod "8e7ccc35-38ad-4677-9cf3-305bc0d322c4" (UID: "8e7ccc35-38ad-4677-9cf3-305bc0d322c4"). InnerVolumeSpecName "kube-api-access-q5dfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.934802 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e7ccc35-38ad-4677-9cf3-305bc0d322c4" (UID: "8e7ccc35-38ad-4677-9cf3-305bc0d322c4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.938334 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-config" (OuterVolumeSpecName: "config") pod "8e7ccc35-38ad-4677-9cf3-305bc0d322c4" (UID: "8e7ccc35-38ad-4677-9cf3-305bc0d322c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.939700 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e7ccc35-38ad-4677-9cf3-305bc0d322c4" (UID: "8e7ccc35-38ad-4677-9cf3-305bc0d322c4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.969485 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.969515 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.969524 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5dfs\" (UniqueName: \"kubernetes.io/projected/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-kube-api-access-q5dfs\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.969537 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:58 crc kubenswrapper[4715]: I1204 14:22:58.976396 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e7ccc35-38ad-4677-9cf3-305bc0d322c4" (UID: "8e7ccc35-38ad-4677-9cf3-305bc0d322c4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.023429 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e7ccc35-38ad-4677-9cf3-305bc0d322c4" (UID: "8e7ccc35-38ad-4677-9cf3-305bc0d322c4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.075501 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.075545 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e7ccc35-38ad-4677-9cf3-305bc0d322c4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.179948 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-rbrt6"] Dec 04 14:22:59 crc kubenswrapper[4715]: W1204 14:22:59.192282 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfe175eb_a594_4a23_a057_b3cbc53cad3e.slice/crio-d85855a47774bd812822b546820b296a041b83587d4bc22ddc3508872230362e WatchSource:0}: Error finding container d85855a47774bd812822b546820b296a041b83587d4bc22ddc3508872230362e: Status 404 returned error can't find the container with id d85855a47774bd812822b546820b296a041b83587d4bc22ddc3508872230362e Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.412697 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" event={"ID":"8e7ccc35-38ad-4677-9cf3-305bc0d322c4","Type":"ContainerDied","Data":"04252d6a21d8ff9855a3c8e22aad6ce308b5fd0885e810797e14ce80786fe7a4"} Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.413019 4715 scope.go:117] "RemoveContainer" containerID="5c2ecfd532559c3fe36e2edcdb4dc3f60982356d9bd62a904f2e04d27f78619f" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.413061 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4wwbd" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.421283 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" event={"ID":"dfe175eb-a594-4a23-a057-b3cbc53cad3e","Type":"ContainerStarted","Data":"d85855a47774bd812822b546820b296a041b83587d4bc22ddc3508872230362e"} Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.482694 4715 scope.go:117] "RemoveContainer" containerID="39feee6dbda3f2962e579ed067fb5942d8ec558e88cd78dfba25851f88627f5c" Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.491188 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4wwbd"] Dec 04 14:22:59 crc kubenswrapper[4715]: I1204 14:22:59.502298 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4wwbd"] Dec 04 14:23:00 crc kubenswrapper[4715]: I1204 14:23:00.435085 4715 generic.go:334] "Generic (PLEG): container finished" podID="dfe175eb-a594-4a23-a057-b3cbc53cad3e" containerID="e615bb0cfba8dae5fb4a5dd600823ed9cfc5359dcf185897c3eff9bbaa69ab37" exitCode=0 Dec 04 14:23:00 crc kubenswrapper[4715]: I1204 14:23:00.435151 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" event={"ID":"dfe175eb-a594-4a23-a057-b3cbc53cad3e","Type":"ContainerDied","Data":"e615bb0cfba8dae5fb4a5dd600823ed9cfc5359dcf185897c3eff9bbaa69ab37"} Dec 04 14:23:01 crc kubenswrapper[4715]: I1204 14:23:01.195055 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" path="/var/lib/kubelet/pods/8e7ccc35-38ad-4677-9cf3-305bc0d322c4/volumes" Dec 04 14:23:01 crc kubenswrapper[4715]: I1204 14:23:01.464867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" event={"ID":"dfe175eb-a594-4a23-a057-b3cbc53cad3e","Type":"ContainerStarted","Data":"8f3cb1578b3a7bc0d6b19c6a39ecab23cd7de100b820edb12425541303992743"} Dec 04 14:23:01 crc kubenswrapper[4715]: I1204 14:23:01.466247 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:23:01 crc kubenswrapper[4715]: I1204 14:23:01.491450 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" podStartSLOduration=3.4914306330000002 podStartE2EDuration="3.491430633s" podCreationTimestamp="2025-12-04 14:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:23:01.486704956 +0000 UTC m=+1538.555423171" watchObservedRunningTime="2025-12-04 14:23:01.491430633 +0000 UTC m=+1538.560148848" Dec 04 14:23:08 crc kubenswrapper[4715]: I1204 14:23:08.625113 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-rbrt6" Dec 04 14:23:08 crc kubenswrapper[4715]: I1204 14:23:08.684746 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-l4m5p"] Dec 04 14:23:08 crc kubenswrapper[4715]: I1204 14:23:08.685068 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerName="dnsmasq-dns" containerID="cri-o://173c01970ca25a60a0b38a2bdbb0a44ea195aaea53077bc68c8a8de757b3b550" gracePeriod=10 Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.548978 4715 generic.go:334] "Generic (PLEG): container finished" podID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerID="173c01970ca25a60a0b38a2bdbb0a44ea195aaea53077bc68c8a8de757b3b550" exitCode=0 Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.549509 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" event={"ID":"d7b82945-5f57-43bb-bb84-6d112339d87f","Type":"ContainerDied","Data":"173c01970ca25a60a0b38a2bdbb0a44ea195aaea53077bc68c8a8de757b3b550"} Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.690168 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854084 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-sb\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854174 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-openstack-edpm-ipam\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854204 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psnxx\" (UniqueName: \"kubernetes.io/projected/d7b82945-5f57-43bb-bb84-6d112339d87f-kube-api-access-psnxx\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854330 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-nb\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854379 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-config\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854486 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-svc\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.854573 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-swift-storage-0\") pod \"d7b82945-5f57-43bb-bb84-6d112339d87f\" (UID: \"d7b82945-5f57-43bb-bb84-6d112339d87f\") " Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.866310 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b82945-5f57-43bb-bb84-6d112339d87f-kube-api-access-psnxx" (OuterVolumeSpecName: "kube-api-access-psnxx") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "kube-api-access-psnxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.907654 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.918817 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.921127 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-config" (OuterVolumeSpecName: "config") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.922116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.923563 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.930845 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7b82945-5f57-43bb-bb84-6d112339d87f" (UID: "d7b82945-5f57-43bb-bb84-6d112339d87f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956594 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956632 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956641 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956651 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956659 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956667 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d7b82945-5f57-43bb-bb84-6d112339d87f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:09 crc kubenswrapper[4715]: I1204 14:23:09.956686 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psnxx\" (UniqueName: \"kubernetes.io/projected/d7b82945-5f57-43bb-bb84-6d112339d87f-kube-api-access-psnxx\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:10 crc kubenswrapper[4715]: I1204 14:23:10.685634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" event={"ID":"d7b82945-5f57-43bb-bb84-6d112339d87f","Type":"ContainerDied","Data":"55b15e4cc4588ed49e2bcdd26ad89899467312c3bcd89c88463481b0eb8f19a0"} Dec 04 14:23:10 crc kubenswrapper[4715]: I1204 14:23:10.687599 4715 scope.go:117] "RemoveContainer" containerID="173c01970ca25a60a0b38a2bdbb0a44ea195aaea53077bc68c8a8de757b3b550" Dec 04 14:23:10 crc kubenswrapper[4715]: I1204 14:23:10.685695 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-l4m5p" Dec 04 14:23:10 crc kubenswrapper[4715]: I1204 14:23:10.720184 4715 scope.go:117] "RemoveContainer" containerID="46b21ea32285b0757d68c0f98b53b4b7083d2c108158bb922bc283eb1a7910d0" Dec 04 14:23:10 crc kubenswrapper[4715]: I1204 14:23:10.726126 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-l4m5p"] Dec 04 14:23:10 crc kubenswrapper[4715]: I1204 14:23:10.739058 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-l4m5p"] Dec 04 14:23:11 crc kubenswrapper[4715]: I1204 14:23:11.192462 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" path="/var/lib/kubelet/pods/d7b82945-5f57-43bb-bb84-6d112339d87f/volumes" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.177959 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt"] Dec 04 14:23:21 crc kubenswrapper[4715]: E1204 14:23:21.179137 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerName="dnsmasq-dns" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.179157 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerName="dnsmasq-dns" Dec 04 14:23:21 crc kubenswrapper[4715]: E1204 14:23:21.179211 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerName="init" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.179219 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerName="init" Dec 04 14:23:21 crc kubenswrapper[4715]: E1204 14:23:21.179237 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerName="init" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.179244 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerName="init" Dec 04 14:23:21 crc kubenswrapper[4715]: E1204 14:23:21.179257 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerName="dnsmasq-dns" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.179265 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerName="dnsmasq-dns" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.179504 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b82945-5f57-43bb-bb84-6d112339d87f" containerName="dnsmasq-dns" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.179520 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e7ccc35-38ad-4677-9cf3-305bc0d322c4" containerName="dnsmasq-dns" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.180586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.182972 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.183661 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.183958 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.193403 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt"] Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.202109 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.227932 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.228061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.228123 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.228201 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7skzw\" (UniqueName: \"kubernetes.io/projected/76dfdae9-34be-48e3-a220-4322d19e7260-kube-api-access-7skzw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.328976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.329065 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.329115 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.329141 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7skzw\" (UniqueName: \"kubernetes.io/projected/76dfdae9-34be-48e3-a220-4322d19e7260-kube-api-access-7skzw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.336489 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.337708 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.340018 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.348386 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7skzw\" (UniqueName: \"kubernetes.io/projected/76dfdae9-34be-48e3-a220-4322d19e7260-kube-api-access-7skzw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:21 crc kubenswrapper[4715]: I1204 14:23:21.511322 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:22 crc kubenswrapper[4715]: I1204 14:23:22.038939 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt"] Dec 04 14:23:22 crc kubenswrapper[4715]: I1204 14:23:22.906005 4715 generic.go:334] "Generic (PLEG): container finished" podID="72cfc89d-7970-45e9-be5f-3f5f5d3f520d" containerID="db4e68667a1ccefd7cac3522c126dcbb29b5c97e9397f0a348af19c78a408075" exitCode=0 Dec 04 14:23:22 crc kubenswrapper[4715]: I1204 14:23:22.906067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"72cfc89d-7970-45e9-be5f-3f5f5d3f520d","Type":"ContainerDied","Data":"db4e68667a1ccefd7cac3522c126dcbb29b5c97e9397f0a348af19c78a408075"} Dec 04 14:23:22 crc kubenswrapper[4715]: I1204 14:23:22.911453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" event={"ID":"76dfdae9-34be-48e3-a220-4322d19e7260","Type":"ContainerStarted","Data":"94e57dd44e4971444d31fd6460923feb3473897bead220ca4fd4dd8c6c9ce5f1"} Dec 04 14:23:23 crc kubenswrapper[4715]: I1204 14:23:23.923495 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"72cfc89d-7970-45e9-be5f-3f5f5d3f520d","Type":"ContainerStarted","Data":"5f5cd8c00071fea80231acf8db6a042b614410c4d10715cf79b80a873e741509"} Dec 04 14:23:23 crc kubenswrapper[4715]: I1204 14:23:23.923978 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 14:23:23 crc kubenswrapper[4715]: I1204 14:23:23.953723 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.953696495 podStartE2EDuration="36.953696495s" podCreationTimestamp="2025-12-04 14:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:23:23.95051393 +0000 UTC m=+1561.019232175" watchObservedRunningTime="2025-12-04 14:23:23.953696495 +0000 UTC m=+1561.022414710" Dec 04 14:23:24 crc kubenswrapper[4715]: I1204 14:23:24.941319 4715 generic.go:334] "Generic (PLEG): container finished" podID="45d230ae-3ae5-4d87-8979-bbd1fd8651cd" containerID="f9e93287d1381b1cee1ec7dc34dac764b31d9bf7e7f2d0ff0d4517f97eefce2a" exitCode=0 Dec 04 14:23:24 crc kubenswrapper[4715]: I1204 14:23:24.943427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45d230ae-3ae5-4d87-8979-bbd1fd8651cd","Type":"ContainerDied","Data":"f9e93287d1381b1cee1ec7dc34dac764b31d9bf7e7f2d0ff0d4517f97eefce2a"} Dec 04 14:23:25 crc kubenswrapper[4715]: I1204 14:23:25.958272 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45d230ae-3ae5-4d87-8979-bbd1fd8651cd","Type":"ContainerStarted","Data":"6b600f91843c512fe5fed6cb0c10cda3f4896ebac02657c2456cb9771260714c"} Dec 04 14:23:25 crc kubenswrapper[4715]: I1204 14:23:25.960519 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:23:25 crc kubenswrapper[4715]: I1204 14:23:25.995195 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.995176427 podStartE2EDuration="37.995176427s" podCreationTimestamp="2025-12-04 14:22:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:23:25.979967679 +0000 UTC m=+1563.048685894" watchObservedRunningTime="2025-12-04 14:23:25.995176427 +0000 UTC m=+1563.063894642" Dec 04 14:23:31 crc kubenswrapper[4715]: I1204 14:23:31.681184 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:23:32 crc kubenswrapper[4715]: I1204 14:23:32.030688 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" event={"ID":"76dfdae9-34be-48e3-a220-4322d19e7260","Type":"ContainerStarted","Data":"235cbe21e341fa604f369f1ea0bb08b56d721d6be8a24b2644daa07eba8368df"} Dec 04 14:23:32 crc kubenswrapper[4715]: I1204 14:23:32.053460 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" podStartSLOduration=1.421011969 podStartE2EDuration="11.053435487s" podCreationTimestamp="2025-12-04 14:23:21 +0000 UTC" firstStartedPulling="2025-12-04 14:23:22.046337013 +0000 UTC m=+1559.115055228" lastFinishedPulling="2025-12-04 14:23:31.678760531 +0000 UTC m=+1568.747478746" observedRunningTime="2025-12-04 14:23:32.047002854 +0000 UTC m=+1569.115721069" watchObservedRunningTime="2025-12-04 14:23:32.053435487 +0000 UTC m=+1569.122153712" Dec 04 14:23:36 crc kubenswrapper[4715]: I1204 14:23:36.451687 4715 scope.go:117] "RemoveContainer" containerID="d12875433248aea72fba337a5e13556dd5b49728876f9762caae11b9604f521d" Dec 04 14:23:36 crc kubenswrapper[4715]: I1204 14:23:36.476965 4715 scope.go:117] "RemoveContainer" containerID="a350c8b672df37f357e4f0ad7859a13a7ec729f5524dc3e5406cf053fd77508c" Dec 04 14:23:37 crc kubenswrapper[4715]: I1204 14:23:37.674244 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 14:23:39 crc kubenswrapper[4715]: I1204 14:23:39.163316 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:23:43 crc kubenswrapper[4715]: I1204 14:23:43.148245 4715 generic.go:334] "Generic (PLEG): container finished" podID="76dfdae9-34be-48e3-a220-4322d19e7260" containerID="235cbe21e341fa604f369f1ea0bb08b56d721d6be8a24b2644daa07eba8368df" exitCode=0 Dec 04 14:23:43 crc kubenswrapper[4715]: I1204 14:23:43.149067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" event={"ID":"76dfdae9-34be-48e3-a220-4322d19e7260","Type":"ContainerDied","Data":"235cbe21e341fa604f369f1ea0bb08b56d721d6be8a24b2644daa07eba8368df"} Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.620136 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.756697 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-repo-setup-combined-ca-bundle\") pod \"76dfdae9-34be-48e3-a220-4322d19e7260\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.757091 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7skzw\" (UniqueName: \"kubernetes.io/projected/76dfdae9-34be-48e3-a220-4322d19e7260-kube-api-access-7skzw\") pod \"76dfdae9-34be-48e3-a220-4322d19e7260\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.757126 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-inventory\") pod \"76dfdae9-34be-48e3-a220-4322d19e7260\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.757184 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-ssh-key\") pod \"76dfdae9-34be-48e3-a220-4322d19e7260\" (UID: \"76dfdae9-34be-48e3-a220-4322d19e7260\") " Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.763190 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76dfdae9-34be-48e3-a220-4322d19e7260-kube-api-access-7skzw" (OuterVolumeSpecName: "kube-api-access-7skzw") pod "76dfdae9-34be-48e3-a220-4322d19e7260" (UID: "76dfdae9-34be-48e3-a220-4322d19e7260"). InnerVolumeSpecName "kube-api-access-7skzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.764596 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "76dfdae9-34be-48e3-a220-4322d19e7260" (UID: "76dfdae9-34be-48e3-a220-4322d19e7260"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.785641 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "76dfdae9-34be-48e3-a220-4322d19e7260" (UID: "76dfdae9-34be-48e3-a220-4322d19e7260"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.797412 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-inventory" (OuterVolumeSpecName: "inventory") pod "76dfdae9-34be-48e3-a220-4322d19e7260" (UID: "76dfdae9-34be-48e3-a220-4322d19e7260"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.861201 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7skzw\" (UniqueName: \"kubernetes.io/projected/76dfdae9-34be-48e3-a220-4322d19e7260-kube-api-access-7skzw\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.861283 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.861298 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:44 crc kubenswrapper[4715]: I1204 14:23:44.861321 4715 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dfdae9-34be-48e3-a220-4322d19e7260-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.172907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" event={"ID":"76dfdae9-34be-48e3-a220-4322d19e7260","Type":"ContainerDied","Data":"94e57dd44e4971444d31fd6460923feb3473897bead220ca4fd4dd8c6c9ce5f1"} Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.173253 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94e57dd44e4971444d31fd6460923feb3473897bead220ca4fd4dd8c6c9ce5f1" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.173079 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.292804 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght"] Dec 04 14:23:45 crc kubenswrapper[4715]: E1204 14:23:45.293272 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76dfdae9-34be-48e3-a220-4322d19e7260" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.293299 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="76dfdae9-34be-48e3-a220-4322d19e7260" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.293547 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="76dfdae9-34be-48e3-a220-4322d19e7260" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.294430 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.297435 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.297634 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.297702 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.297703 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.307155 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght"] Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.473535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.473611 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkfn9\" (UniqueName: \"kubernetes.io/projected/9542e2ff-8723-4d1d-b17c-277976018a11-kube-api-access-kkfn9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.473684 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.575641 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkfn9\" (UniqueName: \"kubernetes.io/projected/9542e2ff-8723-4d1d-b17c-277976018a11-kube-api-access-kkfn9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.576156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.576826 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.581903 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.595703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.621954 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkfn9\" (UniqueName: \"kubernetes.io/projected/9542e2ff-8723-4d1d-b17c-277976018a11-kube-api-access-kkfn9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-plght\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:45 crc kubenswrapper[4715]: I1204 14:23:45.922274 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:46 crc kubenswrapper[4715]: I1204 14:23:46.495609 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght"] Dec 04 14:23:47 crc kubenswrapper[4715]: I1204 14:23:47.194120 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" event={"ID":"9542e2ff-8723-4d1d-b17c-277976018a11","Type":"ContainerStarted","Data":"82905acc6dffb9953270de1944686eb04f94ffb2c5cbc54f2b07551f3721e0ab"} Dec 04 14:23:48 crc kubenswrapper[4715]: I1204 14:23:48.205576 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" event={"ID":"9542e2ff-8723-4d1d-b17c-277976018a11","Type":"ContainerStarted","Data":"659c470b66fe2351ee149ad668969d60c1b108a6d1012469d67dc85a7f40b0ae"} Dec 04 14:23:48 crc kubenswrapper[4715]: I1204 14:23:48.229555 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" podStartSLOduration=2.765435676 podStartE2EDuration="3.229531692s" podCreationTimestamp="2025-12-04 14:23:45 +0000 UTC" firstStartedPulling="2025-12-04 14:23:46.499277813 +0000 UTC m=+1583.567996028" lastFinishedPulling="2025-12-04 14:23:46.963373829 +0000 UTC m=+1584.032092044" observedRunningTime="2025-12-04 14:23:48.222825382 +0000 UTC m=+1585.291543597" watchObservedRunningTime="2025-12-04 14:23:48.229531692 +0000 UTC m=+1585.298249907" Dec 04 14:23:50 crc kubenswrapper[4715]: I1204 14:23:50.228566 4715 generic.go:334] "Generic (PLEG): container finished" podID="9542e2ff-8723-4d1d-b17c-277976018a11" containerID="659c470b66fe2351ee149ad668969d60c1b108a6d1012469d67dc85a7f40b0ae" exitCode=0 Dec 04 14:23:50 crc kubenswrapper[4715]: I1204 14:23:50.228768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" event={"ID":"9542e2ff-8723-4d1d-b17c-277976018a11","Type":"ContainerDied","Data":"659c470b66fe2351ee149ad668969d60c1b108a6d1012469d67dc85a7f40b0ae"} Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.714258 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.801026 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-ssh-key\") pod \"9542e2ff-8723-4d1d-b17c-277976018a11\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.801198 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-inventory\") pod \"9542e2ff-8723-4d1d-b17c-277976018a11\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.801240 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkfn9\" (UniqueName: \"kubernetes.io/projected/9542e2ff-8723-4d1d-b17c-277976018a11-kube-api-access-kkfn9\") pod \"9542e2ff-8723-4d1d-b17c-277976018a11\" (UID: \"9542e2ff-8723-4d1d-b17c-277976018a11\") " Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.809263 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9542e2ff-8723-4d1d-b17c-277976018a11-kube-api-access-kkfn9" (OuterVolumeSpecName: "kube-api-access-kkfn9") pod "9542e2ff-8723-4d1d-b17c-277976018a11" (UID: "9542e2ff-8723-4d1d-b17c-277976018a11"). InnerVolumeSpecName "kube-api-access-kkfn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.831203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9542e2ff-8723-4d1d-b17c-277976018a11" (UID: "9542e2ff-8723-4d1d-b17c-277976018a11"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.834538 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-inventory" (OuterVolumeSpecName: "inventory") pod "9542e2ff-8723-4d1d-b17c-277976018a11" (UID: "9542e2ff-8723-4d1d-b17c-277976018a11"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.903534 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.903573 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkfn9\" (UniqueName: \"kubernetes.io/projected/9542e2ff-8723-4d1d-b17c-277976018a11-kube-api-access-kkfn9\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:51 crc kubenswrapper[4715]: I1204 14:23:51.903585 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9542e2ff-8723-4d1d-b17c-277976018a11-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.251147 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" event={"ID":"9542e2ff-8723-4d1d-b17c-277976018a11","Type":"ContainerDied","Data":"82905acc6dffb9953270de1944686eb04f94ffb2c5cbc54f2b07551f3721e0ab"} Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.251188 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82905acc6dffb9953270de1944686eb04f94ffb2c5cbc54f2b07551f3721e0ab" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.251252 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-plght" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.363083 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt"] Dec 04 14:23:52 crc kubenswrapper[4715]: E1204 14:23:52.364175 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9542e2ff-8723-4d1d-b17c-277976018a11" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.364213 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9542e2ff-8723-4d1d-b17c-277976018a11" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.364553 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9542e2ff-8723-4d1d-b17c-277976018a11" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.367423 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.375919 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.376628 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.377062 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.377295 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.414783 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt"] Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.520094 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tqmq\" (UniqueName: \"kubernetes.io/projected/b6a57886-292b-4dac-bfb6-a51971fcde74-kube-api-access-2tqmq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.520329 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.520366 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.520396 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.622804 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.622879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.622910 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.622957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tqmq\" (UniqueName: \"kubernetes.io/projected/b6a57886-292b-4dac-bfb6-a51971fcde74-kube-api-access-2tqmq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.628805 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.635831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.637881 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.644141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tqmq\" (UniqueName: \"kubernetes.io/projected/b6a57886-292b-4dac-bfb6-a51971fcde74-kube-api-access-2tqmq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:52 crc kubenswrapper[4715]: I1204 14:23:52.705489 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:23:53 crc kubenswrapper[4715]: I1204 14:23:53.223991 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt"] Dec 04 14:23:53 crc kubenswrapper[4715]: I1204 14:23:53.262307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" event={"ID":"b6a57886-292b-4dac-bfb6-a51971fcde74","Type":"ContainerStarted","Data":"548007b949de621255224fdc51d7419d1b83b820aac5ee9f809845210538c5cf"} Dec 04 14:23:55 crc kubenswrapper[4715]: I1204 14:23:55.285879 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" event={"ID":"b6a57886-292b-4dac-bfb6-a51971fcde74","Type":"ContainerStarted","Data":"c7ae6e0fc9d646b5bcd00aed698b108813cc171812c33d5958054e8ed1b53c65"} Dec 04 14:23:55 crc kubenswrapper[4715]: I1204 14:23:55.305116 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" podStartSLOduration=1.868975351 podStartE2EDuration="3.305097126s" podCreationTimestamp="2025-12-04 14:23:52 +0000 UTC" firstStartedPulling="2025-12-04 14:23:53.223680312 +0000 UTC m=+1590.292398527" lastFinishedPulling="2025-12-04 14:23:54.659802087 +0000 UTC m=+1591.728520302" observedRunningTime="2025-12-04 14:23:55.302929708 +0000 UTC m=+1592.371647923" watchObservedRunningTime="2025-12-04 14:23:55.305097126 +0000 UTC m=+1592.373815341" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.402153 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9kmfn"] Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.410395 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.417923 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kmfn"] Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.463780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whkwl\" (UniqueName: \"kubernetes.io/projected/1d778c43-2671-42ae-a00b-67a188e54a9b-kube-api-access-whkwl\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.463854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-catalog-content\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.463903 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-utilities\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.565625 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-catalog-content\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.566401 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-utilities\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.566416 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-catalog-content\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.566459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-utilities\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.566707 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whkwl\" (UniqueName: \"kubernetes.io/projected/1d778c43-2671-42ae-a00b-67a188e54a9b-kube-api-access-whkwl\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.594944 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whkwl\" (UniqueName: \"kubernetes.io/projected/1d778c43-2671-42ae-a00b-67a188e54a9b-kube-api-access-whkwl\") pod \"redhat-marketplace-9kmfn\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:04 crc kubenswrapper[4715]: I1204 14:24:04.743870 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:05 crc kubenswrapper[4715]: I1204 14:24:05.268324 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kmfn"] Dec 04 14:24:05 crc kubenswrapper[4715]: I1204 14:24:05.389632 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kmfn" event={"ID":"1d778c43-2671-42ae-a00b-67a188e54a9b","Type":"ContainerStarted","Data":"e66db3c435c3b49581978d7249442036c8b5f90a5b4c5fdb64c55f6e661efcdc"} Dec 04 14:24:06 crc kubenswrapper[4715]: I1204 14:24:06.405724 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerID="a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0" exitCode=0 Dec 04 14:24:06 crc kubenswrapper[4715]: I1204 14:24:06.405795 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kmfn" event={"ID":"1d778c43-2671-42ae-a00b-67a188e54a9b","Type":"ContainerDied","Data":"a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0"} Dec 04 14:24:08 crc kubenswrapper[4715]: I1204 14:24:08.429766 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerID="eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143" exitCode=0 Dec 04 14:24:08 crc kubenswrapper[4715]: I1204 14:24:08.429883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kmfn" event={"ID":"1d778c43-2671-42ae-a00b-67a188e54a9b","Type":"ContainerDied","Data":"eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143"} Dec 04 14:24:09 crc kubenswrapper[4715]: I1204 14:24:09.442882 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kmfn" event={"ID":"1d778c43-2671-42ae-a00b-67a188e54a9b","Type":"ContainerStarted","Data":"86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb"} Dec 04 14:24:09 crc kubenswrapper[4715]: I1204 14:24:09.467576 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9kmfn" podStartSLOduration=2.982381297 podStartE2EDuration="5.467558897s" podCreationTimestamp="2025-12-04 14:24:04 +0000 UTC" firstStartedPulling="2025-12-04 14:24:06.408141905 +0000 UTC m=+1603.476860120" lastFinishedPulling="2025-12-04 14:24:08.893319475 +0000 UTC m=+1605.962037720" observedRunningTime="2025-12-04 14:24:09.464115035 +0000 UTC m=+1606.532833260" watchObservedRunningTime="2025-12-04 14:24:09.467558897 +0000 UTC m=+1606.536277112" Dec 04 14:24:14 crc kubenswrapper[4715]: I1204 14:24:14.744385 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:14 crc kubenswrapper[4715]: I1204 14:24:14.745145 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:14 crc kubenswrapper[4715]: I1204 14:24:14.803435 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:15 crc kubenswrapper[4715]: I1204 14:24:15.543648 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:15 crc kubenswrapper[4715]: I1204 14:24:15.595911 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kmfn"] Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.452293 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9fsnm"] Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.455325 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.466774 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fsnm"] Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.513600 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9kmfn" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="registry-server" containerID="cri-o://86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb" gracePeriod=2 Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.560423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpq7l\" (UniqueName: \"kubernetes.io/projected/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-kube-api-access-tpq7l\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.560556 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-utilities\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.560816 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-catalog-content\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.667765 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpq7l\" (UniqueName: \"kubernetes.io/projected/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-kube-api-access-tpq7l\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.668367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-utilities\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.668491 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-catalog-content\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.669353 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-utilities\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.669535 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-catalog-content\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.697531 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpq7l\" (UniqueName: \"kubernetes.io/projected/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-kube-api-access-tpq7l\") pod \"community-operators-9fsnm\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:17 crc kubenswrapper[4715]: I1204 14:24:17.788129 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.161014 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.287901 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-catalog-content\") pod \"1d778c43-2671-42ae-a00b-67a188e54a9b\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.287991 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whkwl\" (UniqueName: \"kubernetes.io/projected/1d778c43-2671-42ae-a00b-67a188e54a9b-kube-api-access-whkwl\") pod \"1d778c43-2671-42ae-a00b-67a188e54a9b\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.288047 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-utilities\") pod \"1d778c43-2671-42ae-a00b-67a188e54a9b\" (UID: \"1d778c43-2671-42ae-a00b-67a188e54a9b\") " Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.290666 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-utilities" (OuterVolumeSpecName: "utilities") pod "1d778c43-2671-42ae-a00b-67a188e54a9b" (UID: "1d778c43-2671-42ae-a00b-67a188e54a9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.295346 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d778c43-2671-42ae-a00b-67a188e54a9b-kube-api-access-whkwl" (OuterVolumeSpecName: "kube-api-access-whkwl") pod "1d778c43-2671-42ae-a00b-67a188e54a9b" (UID: "1d778c43-2671-42ae-a00b-67a188e54a9b"). InnerVolumeSpecName "kube-api-access-whkwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.302486 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d778c43-2671-42ae-a00b-67a188e54a9b" (UID: "1d778c43-2671-42ae-a00b-67a188e54a9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.390825 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.390875 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whkwl\" (UniqueName: \"kubernetes.io/projected/1d778c43-2671-42ae-a00b-67a188e54a9b-kube-api-access-whkwl\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.390894 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d778c43-2671-42ae-a00b-67a188e54a9b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.497607 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fsnm"] Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.524904 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerStarted","Data":"922a94ed5febc78b91e90eb5f21f50c0ddc40b437738fbcf425dc0eddc445268"} Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.528472 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerID="86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb" exitCode=0 Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.528506 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kmfn" event={"ID":"1d778c43-2671-42ae-a00b-67a188e54a9b","Type":"ContainerDied","Data":"86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb"} Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.528524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kmfn" event={"ID":"1d778c43-2671-42ae-a00b-67a188e54a9b","Type":"ContainerDied","Data":"e66db3c435c3b49581978d7249442036c8b5f90a5b4c5fdb64c55f6e661efcdc"} Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.528546 4715 scope.go:117] "RemoveContainer" containerID="86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.528663 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kmfn" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.552109 4715 scope.go:117] "RemoveContainer" containerID="eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.574502 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kmfn"] Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.586126 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kmfn"] Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.590379 4715 scope.go:117] "RemoveContainer" containerID="a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.622237 4715 scope.go:117] "RemoveContainer" containerID="86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb" Dec 04 14:24:18 crc kubenswrapper[4715]: E1204 14:24:18.623419 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb\": container with ID starting with 86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb not found: ID does not exist" containerID="86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.623473 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb"} err="failed to get container status \"86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb\": rpc error: code = NotFound desc = could not find container \"86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb\": container with ID starting with 86c85e90b51cb53639b70c3296c63aa75d663d692fc62aeac09e54292c27fedb not found: ID does not exist" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.623513 4715 scope.go:117] "RemoveContainer" containerID="eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143" Dec 04 14:24:18 crc kubenswrapper[4715]: E1204 14:24:18.624050 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143\": container with ID starting with eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143 not found: ID does not exist" containerID="eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.624074 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143"} err="failed to get container status \"eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143\": rpc error: code = NotFound desc = could not find container \"eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143\": container with ID starting with eb20cfc89ed657b8a09dc6f880f164c385b7f0d60daa8f6b56e5f03f09de4143 not found: ID does not exist" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.624091 4715 scope.go:117] "RemoveContainer" containerID="a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0" Dec 04 14:24:18 crc kubenswrapper[4715]: E1204 14:24:18.624636 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0\": container with ID starting with a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0 not found: ID does not exist" containerID="a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0" Dec 04 14:24:18 crc kubenswrapper[4715]: I1204 14:24:18.624662 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0"} err="failed to get container status \"a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0\": rpc error: code = NotFound desc = could not find container \"a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0\": container with ID starting with a114f9ddc9bd86d6a017dbfbd74bb308ac7b33702f8f6c8f4bf4de6407d8e1b0 not found: ID does not exist" Dec 04 14:24:19 crc kubenswrapper[4715]: I1204 14:24:19.196362 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" path="/var/lib/kubelet/pods/1d778c43-2671-42ae-a00b-67a188e54a9b/volumes" Dec 04 14:24:19 crc kubenswrapper[4715]: I1204 14:24:19.540555 4715 generic.go:334] "Generic (PLEG): container finished" podID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerID="24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9" exitCode=0 Dec 04 14:24:19 crc kubenswrapper[4715]: I1204 14:24:19.540609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerDied","Data":"24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9"} Dec 04 14:24:20 crc kubenswrapper[4715]: I1204 14:24:20.556592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerStarted","Data":"738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132"} Dec 04 14:24:21 crc kubenswrapper[4715]: I1204 14:24:21.569714 4715 generic.go:334] "Generic (PLEG): container finished" podID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerID="738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132" exitCode=0 Dec 04 14:24:21 crc kubenswrapper[4715]: I1204 14:24:21.569877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerDied","Data":"738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132"} Dec 04 14:24:22 crc kubenswrapper[4715]: I1204 14:24:22.584161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerStarted","Data":"b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0"} Dec 04 14:24:27 crc kubenswrapper[4715]: I1204 14:24:27.788950 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:27 crc kubenswrapper[4715]: I1204 14:24:27.789521 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:27 crc kubenswrapper[4715]: I1204 14:24:27.841881 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:27 crc kubenswrapper[4715]: I1204 14:24:27.870751 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9fsnm" podStartSLOduration=8.436531822 podStartE2EDuration="10.870724125s" podCreationTimestamp="2025-12-04 14:24:17 +0000 UTC" firstStartedPulling="2025-12-04 14:24:19.543422526 +0000 UTC m=+1616.612140741" lastFinishedPulling="2025-12-04 14:24:21.977614829 +0000 UTC m=+1619.046333044" observedRunningTime="2025-12-04 14:24:22.609698813 +0000 UTC m=+1619.678417058" watchObservedRunningTime="2025-12-04 14:24:27.870724125 +0000 UTC m=+1624.939442350" Dec 04 14:24:28 crc kubenswrapper[4715]: I1204 14:24:28.684549 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:28 crc kubenswrapper[4715]: I1204 14:24:28.734441 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fsnm"] Dec 04 14:24:30 crc kubenswrapper[4715]: I1204 14:24:30.655546 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9fsnm" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="registry-server" containerID="cri-o://b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0" gracePeriod=2 Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.189006 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.257401 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-utilities\") pod \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.258209 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-utilities" (OuterVolumeSpecName: "utilities") pod "c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" (UID: "c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.259126 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-catalog-content\") pod \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.259208 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpq7l\" (UniqueName: \"kubernetes.io/projected/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-kube-api-access-tpq7l\") pod \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\" (UID: \"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0\") " Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.260008 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.277505 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-kube-api-access-tpq7l" (OuterVolumeSpecName: "kube-api-access-tpq7l") pod "c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" (UID: "c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0"). InnerVolumeSpecName "kube-api-access-tpq7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.315004 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" (UID: "c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.361762 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.362104 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpq7l\" (UniqueName: \"kubernetes.io/projected/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0-kube-api-access-tpq7l\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.666318 4715 generic.go:334] "Generic (PLEG): container finished" podID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerID="b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0" exitCode=0 Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.666371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerDied","Data":"b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0"} Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.666432 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fsnm" event={"ID":"c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0","Type":"ContainerDied","Data":"922a94ed5febc78b91e90eb5f21f50c0ddc40b437738fbcf425dc0eddc445268"} Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.666455 4715 scope.go:117] "RemoveContainer" containerID="b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.666387 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fsnm" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.695715 4715 scope.go:117] "RemoveContainer" containerID="738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.699890 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fsnm"] Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.709095 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9fsnm"] Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.732405 4715 scope.go:117] "RemoveContainer" containerID="24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.771901 4715 scope.go:117] "RemoveContainer" containerID="b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0" Dec 04 14:24:31 crc kubenswrapper[4715]: E1204 14:24:31.772361 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0\": container with ID starting with b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0 not found: ID does not exist" containerID="b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.772397 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0"} err="failed to get container status \"b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0\": rpc error: code = NotFound desc = could not find container \"b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0\": container with ID starting with b8a5ade68f704575896cef39711edebd6183243b0d954af5e52a763f42c033e0 not found: ID does not exist" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.772417 4715 scope.go:117] "RemoveContainer" containerID="738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132" Dec 04 14:24:31 crc kubenswrapper[4715]: E1204 14:24:31.772669 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132\": container with ID starting with 738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132 not found: ID does not exist" containerID="738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.772701 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132"} err="failed to get container status \"738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132\": rpc error: code = NotFound desc = could not find container \"738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132\": container with ID starting with 738c251087fe7c3904e6dfcdd570cf15e802e5e5bc85ade1f8f5d969e34a0132 not found: ID does not exist" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.772719 4715 scope.go:117] "RemoveContainer" containerID="24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9" Dec 04 14:24:31 crc kubenswrapper[4715]: E1204 14:24:31.773248 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9\": container with ID starting with 24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9 not found: ID does not exist" containerID="24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9" Dec 04 14:24:31 crc kubenswrapper[4715]: I1204 14:24:31.773304 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9"} err="failed to get container status \"24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9\": rpc error: code = NotFound desc = could not find container \"24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9\": container with ID starting with 24a3f0e56dfa6e9e45c485196f18eaebf227ca9a4531a9bc1b3c3a78207012d9 not found: ID does not exist" Dec 04 14:24:33 crc kubenswrapper[4715]: I1204 14:24:33.196212 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" path="/var/lib/kubelet/pods/c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0/volumes" Dec 04 14:24:36 crc kubenswrapper[4715]: I1204 14:24:36.630415 4715 scope.go:117] "RemoveContainer" containerID="d54ffff1c0ecf1f8e43275f596d60721c329f39ceab316eaa8dde2aa18974237" Dec 04 14:24:36 crc kubenswrapper[4715]: I1204 14:24:36.683178 4715 scope.go:117] "RemoveContainer" containerID="ab30225d43018b007e016d184e5428ecce321a57c30879f40b3b33c0446f54cd" Dec 04 14:24:38 crc kubenswrapper[4715]: I1204 14:24:38.758258 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:24:38 crc kubenswrapper[4715]: I1204 14:24:38.758883 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:25:08 crc kubenswrapper[4715]: I1204 14:25:08.758226 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:25:08 crc kubenswrapper[4715]: I1204 14:25:08.758642 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.700090 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-md5fs"] Dec 04 14:25:20 crc kubenswrapper[4715]: E1204 14:25:20.700989 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="registry-server" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701003 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="registry-server" Dec 04 14:25:20 crc kubenswrapper[4715]: E1204 14:25:20.701026 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="extract-content" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701048 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="extract-content" Dec 04 14:25:20 crc kubenswrapper[4715]: E1204 14:25:20.701059 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="extract-content" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701066 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="extract-content" Dec 04 14:25:20 crc kubenswrapper[4715]: E1204 14:25:20.701085 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="extract-utilities" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701092 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="extract-utilities" Dec 04 14:25:20 crc kubenswrapper[4715]: E1204 14:25:20.701114 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="extract-utilities" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701120 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="extract-utilities" Dec 04 14:25:20 crc kubenswrapper[4715]: E1204 14:25:20.701143 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="registry-server" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701150 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="registry-server" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701348 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7dde7a4-8eda-4a2d-9ef0-bbec6cef18d0" containerName="registry-server" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.701361 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d778c43-2671-42ae-a00b-67a188e54a9b" containerName="registry-server" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.702790 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.711968 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-md5fs"] Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.902948 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-catalog-content\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.903029 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-utilities\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:20 crc kubenswrapper[4715]: I1204 14:25:20.903674 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkzlw\" (UniqueName: \"kubernetes.io/projected/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-kube-api-access-wkzlw\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.006441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkzlw\" (UniqueName: \"kubernetes.io/projected/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-kube-api-access-wkzlw\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.006656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-catalog-content\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.006712 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-utilities\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.007492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-utilities\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.007533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-catalog-content\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.032526 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkzlw\" (UniqueName: \"kubernetes.io/projected/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-kube-api-access-wkzlw\") pod \"certified-operators-md5fs\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.033069 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:21 crc kubenswrapper[4715]: I1204 14:25:21.620915 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-md5fs"] Dec 04 14:25:22 crc kubenswrapper[4715]: I1204 14:25:22.268350 4715 generic.go:334] "Generic (PLEG): container finished" podID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerID="f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803" exitCode=0 Dec 04 14:25:22 crc kubenswrapper[4715]: I1204 14:25:22.268410 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerDied","Data":"f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803"} Dec 04 14:25:22 crc kubenswrapper[4715]: I1204 14:25:22.268447 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerStarted","Data":"e7eff881b609aecd62cb1c730c0dc893806ce09d9b4ee3c00be331192160580b"} Dec 04 14:25:22 crc kubenswrapper[4715]: I1204 14:25:22.271275 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:25:23 crc kubenswrapper[4715]: I1204 14:25:23.281832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerStarted","Data":"9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0"} Dec 04 14:25:24 crc kubenswrapper[4715]: I1204 14:25:24.292276 4715 generic.go:334] "Generic (PLEG): container finished" podID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerID="9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0" exitCode=0 Dec 04 14:25:24 crc kubenswrapper[4715]: I1204 14:25:24.292373 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerDied","Data":"9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0"} Dec 04 14:25:26 crc kubenswrapper[4715]: I1204 14:25:26.338638 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerStarted","Data":"cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0"} Dec 04 14:25:26 crc kubenswrapper[4715]: I1204 14:25:26.364738 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-md5fs" podStartSLOduration=3.805657062 podStartE2EDuration="6.364715885s" podCreationTimestamp="2025-12-04 14:25:20 +0000 UTC" firstStartedPulling="2025-12-04 14:25:22.270926361 +0000 UTC m=+1679.339644576" lastFinishedPulling="2025-12-04 14:25:24.829985184 +0000 UTC m=+1681.898703399" observedRunningTime="2025-12-04 14:25:26.360675537 +0000 UTC m=+1683.429393762" watchObservedRunningTime="2025-12-04 14:25:26.364715885 +0000 UTC m=+1683.433434100" Dec 04 14:25:31 crc kubenswrapper[4715]: I1204 14:25:31.034414 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:31 crc kubenswrapper[4715]: I1204 14:25:31.034722 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:31 crc kubenswrapper[4715]: I1204 14:25:31.086578 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:31 crc kubenswrapper[4715]: I1204 14:25:31.434456 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:31 crc kubenswrapper[4715]: I1204 14:25:31.496798 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-md5fs"] Dec 04 14:25:33 crc kubenswrapper[4715]: I1204 14:25:33.401258 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-md5fs" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="registry-server" containerID="cri-o://cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0" gracePeriod=2 Dec 04 14:25:33 crc kubenswrapper[4715]: I1204 14:25:33.980337 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.124068 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-catalog-content\") pod \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.124128 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-utilities\") pod \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.124204 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkzlw\" (UniqueName: \"kubernetes.io/projected/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-kube-api-access-wkzlw\") pod \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\" (UID: \"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2\") " Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.125787 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-utilities" (OuterVolumeSpecName: "utilities") pod "8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" (UID: "8449ff06-2b6f-487f-aee7-6f7d59fe9fc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.130979 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-kube-api-access-wkzlw" (OuterVolumeSpecName: "kube-api-access-wkzlw") pod "8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" (UID: "8449ff06-2b6f-487f-aee7-6f7d59fe9fc2"). InnerVolumeSpecName "kube-api-access-wkzlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.177496 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" (UID: "8449ff06-2b6f-487f-aee7-6f7d59fe9fc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.227637 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.227680 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.227695 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkzlw\" (UniqueName: \"kubernetes.io/projected/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2-kube-api-access-wkzlw\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.420631 4715 generic.go:334] "Generic (PLEG): container finished" podID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerID="cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0" exitCode=0 Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.420705 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerDied","Data":"cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0"} Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.420736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md5fs" event={"ID":"8449ff06-2b6f-487f-aee7-6f7d59fe9fc2","Type":"ContainerDied","Data":"e7eff881b609aecd62cb1c730c0dc893806ce09d9b4ee3c00be331192160580b"} Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.420746 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md5fs" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.420764 4715 scope.go:117] "RemoveContainer" containerID="cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.449971 4715 scope.go:117] "RemoveContainer" containerID="9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.478944 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-md5fs"] Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.489918 4715 scope.go:117] "RemoveContainer" containerID="f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.496535 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-md5fs"] Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.523867 4715 scope.go:117] "RemoveContainer" containerID="cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0" Dec 04 14:25:34 crc kubenswrapper[4715]: E1204 14:25:34.524312 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0\": container with ID starting with cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0 not found: ID does not exist" containerID="cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.524350 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0"} err="failed to get container status \"cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0\": rpc error: code = NotFound desc = could not find container \"cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0\": container with ID starting with cd5a927d6dbee8a830ad7f3771072e1b56599fc42fd5cf9651fd51d8791383c0 not found: ID does not exist" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.524374 4715 scope.go:117] "RemoveContainer" containerID="9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0" Dec 04 14:25:34 crc kubenswrapper[4715]: E1204 14:25:34.524929 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0\": container with ID starting with 9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0 not found: ID does not exist" containerID="9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.524983 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0"} err="failed to get container status \"9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0\": rpc error: code = NotFound desc = could not find container \"9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0\": container with ID starting with 9f6e1fb4d194bc2ce90380b58ecbb8d1522b56b875fa9ad033856e4fe2b19ce0 not found: ID does not exist" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.525013 4715 scope.go:117] "RemoveContainer" containerID="f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803" Dec 04 14:25:34 crc kubenswrapper[4715]: E1204 14:25:34.526611 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803\": container with ID starting with f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803 not found: ID does not exist" containerID="f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803" Dec 04 14:25:34 crc kubenswrapper[4715]: I1204 14:25:34.526651 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803"} err="failed to get container status \"f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803\": rpc error: code = NotFound desc = could not find container \"f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803\": container with ID starting with f4c4fe8740b5677517abefaf68d44986beefc87b101a3b38b6ee8923c5159803 not found: ID does not exist" Dec 04 14:25:35 crc kubenswrapper[4715]: I1204 14:25:35.191567 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" path="/var/lib/kubelet/pods/8449ff06-2b6f-487f-aee7-6f7d59fe9fc2/volumes" Dec 04 14:25:36 crc kubenswrapper[4715]: I1204 14:25:36.847914 4715 scope.go:117] "RemoveContainer" containerID="80bb690535d08957f268fb451a9b2608f1e5529fbe003046329fd5dbf9c7373b" Dec 04 14:25:36 crc kubenswrapper[4715]: I1204 14:25:36.869611 4715 scope.go:117] "RemoveContainer" containerID="336e9d52fb70bc3782a98e38285ecf20041d9a92a65f76c6640e39d24dc8d24c" Dec 04 14:25:36 crc kubenswrapper[4715]: I1204 14:25:36.894667 4715 scope.go:117] "RemoveContainer" containerID="d9d5504dd22df7cd3fddb8b08067bef9d4f4ee6732924b53b10b4c34fa598d9e" Dec 04 14:25:37 crc kubenswrapper[4715]: I1204 14:25:37.083762 4715 scope.go:117] "RemoveContainer" containerID="4249d77a08662aebb6db84ab2ea51b79e5a7ea600284207d7c0c9160afb94728" Dec 04 14:25:38 crc kubenswrapper[4715]: I1204 14:25:38.758355 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:25:38 crc kubenswrapper[4715]: I1204 14:25:38.758673 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:25:38 crc kubenswrapper[4715]: I1204 14:25:38.758718 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:25:38 crc kubenswrapper[4715]: I1204 14:25:38.759500 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:25:38 crc kubenswrapper[4715]: I1204 14:25:38.759562 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" gracePeriod=600 Dec 04 14:25:38 crc kubenswrapper[4715]: E1204 14:25:38.882650 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:25:39 crc kubenswrapper[4715]: I1204 14:25:39.473283 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" exitCode=0 Dec 04 14:25:39 crc kubenswrapper[4715]: I1204 14:25:39.474102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b"} Dec 04 14:25:39 crc kubenswrapper[4715]: I1204 14:25:39.474278 4715 scope.go:117] "RemoveContainer" containerID="5f8cf4969e2a565cf9483d0133bd3e156b294d54b874a31409400fd261c14615" Dec 04 14:25:39 crc kubenswrapper[4715]: I1204 14:25:39.474962 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:25:39 crc kubenswrapper[4715]: E1204 14:25:39.475398 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:25:54 crc kubenswrapper[4715]: I1204 14:25:54.180653 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:25:54 crc kubenswrapper[4715]: E1204 14:25:54.181387 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:26:06 crc kubenswrapper[4715]: I1204 14:26:06.180792 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:26:06 crc kubenswrapper[4715]: E1204 14:26:06.181491 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:26:17 crc kubenswrapper[4715]: I1204 14:26:17.180637 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:26:17 crc kubenswrapper[4715]: E1204 14:26:17.181360 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:26:28 crc kubenswrapper[4715]: I1204 14:26:28.181958 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:26:28 crc kubenswrapper[4715]: E1204 14:26:28.182729 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:26:37 crc kubenswrapper[4715]: I1204 14:26:37.166717 4715 scope.go:117] "RemoveContainer" containerID="9591a75565166e919304c0389e032ae1a259b9bf26010ef1316121b6822b7bd5" Dec 04 14:26:37 crc kubenswrapper[4715]: I1204 14:26:37.195342 4715 scope.go:117] "RemoveContainer" containerID="b44351b04a9829963f43cdd221275304338a9a87f7d6d2d1b9dbadd920e0567b" Dec 04 14:26:40 crc kubenswrapper[4715]: I1204 14:26:40.181385 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:26:40 crc kubenswrapper[4715]: E1204 14:26:40.182436 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:26:54 crc kubenswrapper[4715]: I1204 14:26:54.180195 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:26:54 crc kubenswrapper[4715]: E1204 14:26:54.181002 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:27:07 crc kubenswrapper[4715]: I1204 14:27:07.382643 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" event={"ID":"b6a57886-292b-4dac-bfb6-a51971fcde74","Type":"ContainerDied","Data":"c7ae6e0fc9d646b5bcd00aed698b108813cc171812c33d5958054e8ed1b53c65"} Dec 04 14:27:07 crc kubenswrapper[4715]: I1204 14:27:07.382677 4715 generic.go:334] "Generic (PLEG): container finished" podID="b6a57886-292b-4dac-bfb6-a51971fcde74" containerID="c7ae6e0fc9d646b5bcd00aed698b108813cc171812c33d5958054e8ed1b53c65" exitCode=0 Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.180488 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:27:08 crc kubenswrapper[4715]: E1204 14:27:08.180790 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.821222 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.907220 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-inventory\") pod \"b6a57886-292b-4dac-bfb6-a51971fcde74\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.907274 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tqmq\" (UniqueName: \"kubernetes.io/projected/b6a57886-292b-4dac-bfb6-a51971fcde74-kube-api-access-2tqmq\") pod \"b6a57886-292b-4dac-bfb6-a51971fcde74\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.907488 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-bootstrap-combined-ca-bundle\") pod \"b6a57886-292b-4dac-bfb6-a51971fcde74\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.907536 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-ssh-key\") pod \"b6a57886-292b-4dac-bfb6-a51971fcde74\" (UID: \"b6a57886-292b-4dac-bfb6-a51971fcde74\") " Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.915826 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b6a57886-292b-4dac-bfb6-a51971fcde74" (UID: "b6a57886-292b-4dac-bfb6-a51971fcde74"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.915907 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6a57886-292b-4dac-bfb6-a51971fcde74-kube-api-access-2tqmq" (OuterVolumeSpecName: "kube-api-access-2tqmq") pod "b6a57886-292b-4dac-bfb6-a51971fcde74" (UID: "b6a57886-292b-4dac-bfb6-a51971fcde74"). InnerVolumeSpecName "kube-api-access-2tqmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.941809 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-inventory" (OuterVolumeSpecName: "inventory") pod "b6a57886-292b-4dac-bfb6-a51971fcde74" (UID: "b6a57886-292b-4dac-bfb6-a51971fcde74"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:27:08 crc kubenswrapper[4715]: I1204 14:27:08.943476 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6a57886-292b-4dac-bfb6-a51971fcde74" (UID: "b6a57886-292b-4dac-bfb6-a51971fcde74"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.010435 4715 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.010812 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.010831 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a57886-292b-4dac-bfb6-a51971fcde74-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.010845 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tqmq\" (UniqueName: \"kubernetes.io/projected/b6a57886-292b-4dac-bfb6-a51971fcde74-kube-api-access-2tqmq\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.408003 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" event={"ID":"b6a57886-292b-4dac-bfb6-a51971fcde74","Type":"ContainerDied","Data":"548007b949de621255224fdc51d7419d1b83b820aac5ee9f809845210538c5cf"} Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.408071 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="548007b949de621255224fdc51d7419d1b83b820aac5ee9f809845210538c5cf" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.408095 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.540612 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88"] Dec 04 14:27:09 crc kubenswrapper[4715]: E1204 14:27:09.541260 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="extract-content" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.541293 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="extract-content" Dec 04 14:27:09 crc kubenswrapper[4715]: E1204 14:27:09.541332 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="extract-utilities" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.541344 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="extract-utilities" Dec 04 14:27:09 crc kubenswrapper[4715]: E1204 14:27:09.541385 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="registry-server" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.541395 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="registry-server" Dec 04 14:27:09 crc kubenswrapper[4715]: E1204 14:27:09.541413 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6a57886-292b-4dac-bfb6-a51971fcde74" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.541423 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6a57886-292b-4dac-bfb6-a51971fcde74" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.541697 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8449ff06-2b6f-487f-aee7-6f7d59fe9fc2" containerName="registry-server" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.541723 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6a57886-292b-4dac-bfb6-a51971fcde74" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.543077 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.549406 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.549429 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.549497 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.549428 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.557135 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88"] Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.623332 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.623428 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.623468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcn82\" (UniqueName: \"kubernetes.io/projected/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-kube-api-access-dcn82\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.725293 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.725361 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcn82\" (UniqueName: \"kubernetes.io/projected/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-kube-api-access-dcn82\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.725535 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.731695 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.735798 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.743012 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcn82\" (UniqueName: \"kubernetes.io/projected/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-kube-api-access-dcn82\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-8dv88\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:09 crc kubenswrapper[4715]: I1204 14:27:09.883898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:27:10 crc kubenswrapper[4715]: I1204 14:27:10.417001 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88"] Dec 04 14:27:11 crc kubenswrapper[4715]: I1204 14:27:11.432199 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" event={"ID":"a8a8da4d-f578-4428-8b74-9d4efd5f78ee","Type":"ContainerStarted","Data":"d5bf01f43aebeb2d0cd668104163cfa971521df081562b47d8bb51862c1809b7"} Dec 04 14:27:11 crc kubenswrapper[4715]: I1204 14:27:11.436799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" event={"ID":"a8a8da4d-f578-4428-8b74-9d4efd5f78ee","Type":"ContainerStarted","Data":"1fbc264f44db8f9e0dbe2a55b152e4b830aad1ff53616ef78b91a8de367cc6eb"} Dec 04 14:27:11 crc kubenswrapper[4715]: I1204 14:27:11.463309 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" podStartSLOduration=2.040103834 podStartE2EDuration="2.463292842s" podCreationTimestamp="2025-12-04 14:27:09 +0000 UTC" firstStartedPulling="2025-12-04 14:27:10.419052517 +0000 UTC m=+1787.487770732" lastFinishedPulling="2025-12-04 14:27:10.842241525 +0000 UTC m=+1787.910959740" observedRunningTime="2025-12-04 14:27:11.452979544 +0000 UTC m=+1788.521697759" watchObservedRunningTime="2025-12-04 14:27:11.463292842 +0000 UTC m=+1788.532011057" Dec 04 14:27:23 crc kubenswrapper[4715]: I1204 14:27:23.202447 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:27:23 crc kubenswrapper[4715]: E1204 14:27:23.204103 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:27:35 crc kubenswrapper[4715]: I1204 14:27:35.041020 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-74hx8"] Dec 04 14:27:35 crc kubenswrapper[4715]: I1204 14:27:35.050717 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-74hx8"] Dec 04 14:27:35 crc kubenswrapper[4715]: I1204 14:27:35.191221 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8932cfe-78f8-4b68-9312-db06249648aa" path="/var/lib/kubelet/pods/b8932cfe-78f8-4b68-9312-db06249648aa/volumes" Dec 04 14:27:36 crc kubenswrapper[4715]: I1204 14:27:36.030705 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-2988-account-create-update-9spll"] Dec 04 14:27:36 crc kubenswrapper[4715]: I1204 14:27:36.040889 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-2988-account-create-update-9spll"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.036961 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-sfl96"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.045931 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-nc788"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.056478 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-1742-account-create-update-gxqqt"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.069248 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-890e-account-create-update-q5wjc"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.077664 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-nc788"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.085756 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-1742-account-create-update-gxqqt"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.093256 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-sfl96"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.100624 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-890e-account-create-update-q5wjc"] Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.181186 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:27:37 crc kubenswrapper[4715]: E1204 14:27:37.181422 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.195161 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99af1217-29e9-43e2-bb9c-441670a08c75" path="/var/lib/kubelet/pods/99af1217-29e9-43e2-bb9c-441670a08c75/volumes" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.195954 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab09ee34-7f1d-4a57-ac0d-0d809445ba95" path="/var/lib/kubelet/pods/ab09ee34-7f1d-4a57-ac0d-0d809445ba95/volumes" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.196722 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0daee2-9146-4c46-b3f9-8c4008eb6b7b" path="/var/lib/kubelet/pods/af0daee2-9146-4c46-b3f9-8c4008eb6b7b/volumes" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.197464 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b5b100-cb32-4476-9631-74cb35856752" path="/var/lib/kubelet/pods/e7b5b100-cb32-4476-9631-74cb35856752/volumes" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.199193 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f167ccc4-f804-4a6e-b194-0274a320e900" path="/var/lib/kubelet/pods/f167ccc4-f804-4a6e-b194-0274a320e900/volumes" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.270817 4715 scope.go:117] "RemoveContainer" containerID="243b1152e8f12ffcdbfa588c944c85c6389fdb18a26490bb9ca63e2b14687750" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.298881 4715 scope.go:117] "RemoveContainer" containerID="71706ad94f1431d9a3543e36e9034bf7365e382f38af42bf3a493e38a2467ae5" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.349406 4715 scope.go:117] "RemoveContainer" containerID="ff443cd073060401f9a7bb24699b7459b6f14d7447e578263aef790f1d3ff8aa" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.398824 4715 scope.go:117] "RemoveContainer" containerID="fa9754f3f8d690c42db6b5416de8132665cc05d7891cfc44f59dafd7c02379c8" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.442187 4715 scope.go:117] "RemoveContainer" containerID="d0d60bf9da736590699534719c054268068a7018408053753ad7ccd4fed47ba0" Dec 04 14:27:37 crc kubenswrapper[4715]: I1204 14:27:37.486144 4715 scope.go:117] "RemoveContainer" containerID="0accca53065ad2257f5d65487256dc53eedeb76d65ce657e0d238469c9fb2b4d" Dec 04 14:27:48 crc kubenswrapper[4715]: I1204 14:27:48.180834 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:27:48 crc kubenswrapper[4715]: E1204 14:27:48.181726 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:28:01 crc kubenswrapper[4715]: I1204 14:28:01.181533 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:28:01 crc kubenswrapper[4715]: E1204 14:28:01.182310 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:28:06 crc kubenswrapper[4715]: I1204 14:28:06.042673 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-p2wq9"] Dec 04 14:28:06 crc kubenswrapper[4715]: I1204 14:28:06.053764 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-p2wq9"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.037593 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6t92p"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.045696 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b8f3-account-create-update-xllmc"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.053971 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-8j6np"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.062534 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ec3a-account-create-update-p9mp4"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.071462 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6t92p"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.081408 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-8j6np"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.089983 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ec3a-account-create-update-p9mp4"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.098125 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b8f3-account-create-update-xllmc"] Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.191821 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="439588bc-077f-40ac-9b7c-684b1ad07055" path="/var/lib/kubelet/pods/439588bc-077f-40ac-9b7c-684b1ad07055/volumes" Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.192664 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5607a052-6021-469b-a120-b95678e39c43" path="/var/lib/kubelet/pods/5607a052-6021-469b-a120-b95678e39c43/volumes" Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.193546 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8496bf97-de9b-4438-a2d6-2af2046ebdd9" path="/var/lib/kubelet/pods/8496bf97-de9b-4438-a2d6-2af2046ebdd9/volumes" Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.194267 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f42c793-540b-4f39-9325-54858d52bc58" path="/var/lib/kubelet/pods/9f42c793-540b-4f39-9325-54858d52bc58/volumes" Dec 04 14:28:07 crc kubenswrapper[4715]: I1204 14:28:07.195498 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee23d5e5-7936-44fe-8d9a-041a52063147" path="/var/lib/kubelet/pods/ee23d5e5-7936-44fe-8d9a-041a52063147/volumes" Dec 04 14:28:10 crc kubenswrapper[4715]: I1204 14:28:10.028199 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c614-account-create-update-52t56"] Dec 04 14:28:10 crc kubenswrapper[4715]: I1204 14:28:10.036554 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c614-account-create-update-52t56"] Dec 04 14:28:11 crc kubenswrapper[4715]: I1204 14:28:11.196793 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a58c31a-8d3f-4425-bfdf-17c7c5f149aa" path="/var/lib/kubelet/pods/8a58c31a-8d3f-4425-bfdf-17c7c5f149aa/volumes" Dec 04 14:28:12 crc kubenswrapper[4715]: I1204 14:28:12.181416 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:28:12 crc kubenswrapper[4715]: E1204 14:28:12.181720 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:28:23 crc kubenswrapper[4715]: I1204 14:28:23.039929 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gh6ns"] Dec 04 14:28:23 crc kubenswrapper[4715]: I1204 14:28:23.048697 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gh6ns"] Dec 04 14:28:23 crc kubenswrapper[4715]: I1204 14:28:23.195961 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffdd4a77-1d9f-4f90-b604-9182c576913f" path="/var/lib/kubelet/pods/ffdd4a77-1d9f-4f90-b604-9182c576913f/volumes" Dec 04 14:28:24 crc kubenswrapper[4715]: I1204 14:28:24.180488 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:28:24 crc kubenswrapper[4715]: E1204 14:28:24.181244 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:28:28 crc kubenswrapper[4715]: I1204 14:28:28.036239 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ddj77"] Dec 04 14:28:28 crc kubenswrapper[4715]: I1204 14:28:28.046670 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ddj77"] Dec 04 14:28:29 crc kubenswrapper[4715]: I1204 14:28:29.193396 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b59e0b-f6fd-42e0-9780-7c13708c379c" path="/var/lib/kubelet/pods/88b59e0b-f6fd-42e0-9780-7c13708c379c/volumes" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.616332 4715 scope.go:117] "RemoveContainer" containerID="517cc83fdf050232a010ef797a10434f7aecacea989c3043896f3ee79e28f824" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.645895 4715 scope.go:117] "RemoveContainer" containerID="4bfa2f10b9bddae5417d80c0b064c3bf49e30475d9b123ad0e559f8e6b3d6532" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.705608 4715 scope.go:117] "RemoveContainer" containerID="d5ad74cc3099453913a147a19f61db4e1d3624950e0594fb15fb0f42e1711c03" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.763565 4715 scope.go:117] "RemoveContainer" containerID="4155d5f363c8bf805ab45145ec24681cecd4abe1a4dd874b59257211d3d28882" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.804651 4715 scope.go:117] "RemoveContainer" containerID="f0de8923127566d140d82d25dcedf2bcc2a862e2abc3a3863d90fe6f950bf8e3" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.856685 4715 scope.go:117] "RemoveContainer" containerID="a90919ab0e8b4d36775da8dc231de863c7940e1fddf6a502eba1483849e70fc9" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.897668 4715 scope.go:117] "RemoveContainer" containerID="8b6007f3e8883b3289191a9f604bff9e99321a68097772512a143d00483a3292" Dec 04 14:28:37 crc kubenswrapper[4715]: I1204 14:28:37.922781 4715 scope.go:117] "RemoveContainer" containerID="da6c07fa2bd0fbab936b95f437fa6b0b4046b710cddbd71a18cbbcd6caa32ec3" Dec 04 14:28:39 crc kubenswrapper[4715]: I1204 14:28:39.180601 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:28:39 crc kubenswrapper[4715]: E1204 14:28:39.180848 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:28:51 crc kubenswrapper[4715]: I1204 14:28:51.180698 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:28:51 crc kubenswrapper[4715]: E1204 14:28:51.182602 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:29:02 crc kubenswrapper[4715]: I1204 14:29:02.536051 4715 generic.go:334] "Generic (PLEG): container finished" podID="a8a8da4d-f578-4428-8b74-9d4efd5f78ee" containerID="d5bf01f43aebeb2d0cd668104163cfa971521df081562b47d8bb51862c1809b7" exitCode=0 Dec 04 14:29:02 crc kubenswrapper[4715]: I1204 14:29:02.536608 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" event={"ID":"a8a8da4d-f578-4428-8b74-9d4efd5f78ee","Type":"ContainerDied","Data":"d5bf01f43aebeb2d0cd668104163cfa971521df081562b47d8bb51862c1809b7"} Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:03.998385 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.184817 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-inventory\") pod \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.185303 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-ssh-key\") pod \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.185352 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcn82\" (UniqueName: \"kubernetes.io/projected/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-kube-api-access-dcn82\") pod \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\" (UID: \"a8a8da4d-f578-4428-8b74-9d4efd5f78ee\") " Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.203090 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-kube-api-access-dcn82" (OuterVolumeSpecName: "kube-api-access-dcn82") pod "a8a8da4d-f578-4428-8b74-9d4efd5f78ee" (UID: "a8a8da4d-f578-4428-8b74-9d4efd5f78ee"). InnerVolumeSpecName "kube-api-access-dcn82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.219322 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-inventory" (OuterVolumeSpecName: "inventory") pod "a8a8da4d-f578-4428-8b74-9d4efd5f78ee" (UID: "a8a8da4d-f578-4428-8b74-9d4efd5f78ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.220898 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a8a8da4d-f578-4428-8b74-9d4efd5f78ee" (UID: "a8a8da4d-f578-4428-8b74-9d4efd5f78ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.291426 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.291629 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.291686 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcn82\" (UniqueName: \"kubernetes.io/projected/a8a8da4d-f578-4428-8b74-9d4efd5f78ee-kube-api-access-dcn82\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.554532 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" event={"ID":"a8a8da4d-f578-4428-8b74-9d4efd5f78ee","Type":"ContainerDied","Data":"1fbc264f44db8f9e0dbe2a55b152e4b830aad1ff53616ef78b91a8de367cc6eb"} Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.554573 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fbc264f44db8f9e0dbe2a55b152e4b830aad1ff53616ef78b91a8de367cc6eb" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.554595 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-8dv88" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.689381 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w"] Dec 04 14:29:04 crc kubenswrapper[4715]: E1204 14:29:04.689868 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8a8da4d-f578-4428-8b74-9d4efd5f78ee" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.689891 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8a8da4d-f578-4428-8b74-9d4efd5f78ee" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.690148 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8a8da4d-f578-4428-8b74-9d4efd5f78ee" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.690928 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.693789 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.694686 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.701439 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.702283 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.723289 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w"] Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.801313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bfvk\" (UniqueName: \"kubernetes.io/projected/7f4cab97-fa52-47c4-ad8e-7469556526cc-kube-api-access-5bfvk\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.801388 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.801454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.903303 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bfvk\" (UniqueName: \"kubernetes.io/projected/7f4cab97-fa52-47c4-ad8e-7469556526cc-kube-api-access-5bfvk\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.903394 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.903489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.918967 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.918976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:04 crc kubenswrapper[4715]: I1204 14:29:04.926388 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bfvk\" (UniqueName: \"kubernetes.io/projected/7f4cab97-fa52-47c4-ad8e-7469556526cc-kube-api-access-5bfvk\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6t47w\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:05 crc kubenswrapper[4715]: I1204 14:29:05.021764 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:29:05 crc kubenswrapper[4715]: I1204 14:29:05.181272 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:29:05 crc kubenswrapper[4715]: E1204 14:29:05.182016 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:29:05 crc kubenswrapper[4715]: I1204 14:29:05.585973 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w"] Dec 04 14:29:06 crc kubenswrapper[4715]: I1204 14:29:06.573304 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" event={"ID":"7f4cab97-fa52-47c4-ad8e-7469556526cc","Type":"ContainerStarted","Data":"43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57"} Dec 04 14:29:07 crc kubenswrapper[4715]: I1204 14:29:07.584170 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" event={"ID":"7f4cab97-fa52-47c4-ad8e-7469556526cc","Type":"ContainerStarted","Data":"a78b88786d94fb93939e6c72721556388a06ae6b04b4706277fedbe2a6946c4d"} Dec 04 14:29:07 crc kubenswrapper[4715]: I1204 14:29:07.600909 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" podStartSLOduration=2.432435751 podStartE2EDuration="3.600889761s" podCreationTimestamp="2025-12-04 14:29:04 +0000 UTC" firstStartedPulling="2025-12-04 14:29:05.58994347 +0000 UTC m=+1902.658661695" lastFinishedPulling="2025-12-04 14:29:06.75839749 +0000 UTC m=+1903.827115705" observedRunningTime="2025-12-04 14:29:07.599861804 +0000 UTC m=+1904.668580029" watchObservedRunningTime="2025-12-04 14:29:07.600889761 +0000 UTC m=+1904.669607976" Dec 04 14:29:12 crc kubenswrapper[4715]: I1204 14:29:12.042577 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5pp8k"] Dec 04 14:29:12 crc kubenswrapper[4715]: I1204 14:29:12.050639 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5pp8k"] Dec 04 14:29:13 crc kubenswrapper[4715]: I1204 14:29:13.192326 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb889380-79fe-461a-9a87-85c13fd8c716" path="/var/lib/kubelet/pods/fb889380-79fe-461a-9a87-85c13fd8c716/volumes" Dec 04 14:29:16 crc kubenswrapper[4715]: I1204 14:29:16.034321 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-8bbsj"] Dec 04 14:29:16 crc kubenswrapper[4715]: I1204 14:29:16.043254 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-8bbsj"] Dec 04 14:29:17 crc kubenswrapper[4715]: I1204 14:29:17.192342 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ddb5944-20e2-4f19-96c4-ee93ac8cfa02" path="/var/lib/kubelet/pods/2ddb5944-20e2-4f19-96c4-ee93ac8cfa02/volumes" Dec 04 14:29:18 crc kubenswrapper[4715]: I1204 14:29:18.180721 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:29:18 crc kubenswrapper[4715]: E1204 14:29:18.181806 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:29:24 crc kubenswrapper[4715]: I1204 14:29:24.038473 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-s5j8r"] Dec 04 14:29:24 crc kubenswrapper[4715]: I1204 14:29:24.047822 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-s5j8r"] Dec 04 14:29:25 crc kubenswrapper[4715]: I1204 14:29:25.231178 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e311957-d00b-43da-a9dc-ca48386dfa88" path="/var/lib/kubelet/pods/7e311957-d00b-43da-a9dc-ca48386dfa88/volumes" Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.035953 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-cgdwm"] Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.045197 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-5mrnk"] Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.054674 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-5mrnk"] Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.063568 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-cgdwm"] Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.187482 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:29:33 crc kubenswrapper[4715]: E1204 14:29:33.189505 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.192068 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ca3112c-c14e-41af-ad25-c16c7ca1a8c0" path="/var/lib/kubelet/pods/2ca3112c-c14e-41af-ad25-c16c7ca1a8c0/volumes" Dec 04 14:29:33 crc kubenswrapper[4715]: I1204 14:29:33.193318 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ea9b8a-207e-4e7b-9df2-c51ccc497fa1" path="/var/lib/kubelet/pods/95ea9b8a-207e-4e7b-9df2-c51ccc497fa1/volumes" Dec 04 14:29:38 crc kubenswrapper[4715]: I1204 14:29:38.087220 4715 scope.go:117] "RemoveContainer" containerID="2be8db361a9206eb161ca65372c60a8afb28cd4bd530b81bae4438f3887cf052" Dec 04 14:29:38 crc kubenswrapper[4715]: I1204 14:29:38.125250 4715 scope.go:117] "RemoveContainer" containerID="ceb9e9fdc0fbe0e01c1cff1e55fde32653535576aeb494e4e54de204faf484cd" Dec 04 14:29:38 crc kubenswrapper[4715]: I1204 14:29:38.176838 4715 scope.go:117] "RemoveContainer" containerID="96ac05113e2494087f146b6b35619bf4ae4f1f9b50c7c629c76b100a8a2c3180" Dec 04 14:29:38 crc kubenswrapper[4715]: I1204 14:29:38.229196 4715 scope.go:117] "RemoveContainer" containerID="f72345280247cf84a4862946e0d8c5962be25d237d103214868392cdcc6cf55e" Dec 04 14:29:38 crc kubenswrapper[4715]: I1204 14:29:38.261531 4715 scope.go:117] "RemoveContainer" containerID="60669bee47edc974d542a61a67917981b56c1c8b8cb397beeba5befbdbfdda40" Dec 04 14:29:44 crc kubenswrapper[4715]: I1204 14:29:44.180342 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:29:44 crc kubenswrapper[4715]: E1204 14:29:44.181296 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:29:56 crc kubenswrapper[4715]: I1204 14:29:56.180537 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:29:56 crc kubenswrapper[4715]: E1204 14:29:56.182343 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.158948 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h"] Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.162702 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.165839 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.165727 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.172568 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h"] Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.284184 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kns6l\" (UniqueName: \"kubernetes.io/projected/8d505f96-7810-4021-9fd2-5abd46d6ddc6-kube-api-access-kns6l\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.284466 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d505f96-7810-4021-9fd2-5abd46d6ddc6-secret-volume\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.284575 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d505f96-7810-4021-9fd2-5abd46d6ddc6-config-volume\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.386454 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kns6l\" (UniqueName: \"kubernetes.io/projected/8d505f96-7810-4021-9fd2-5abd46d6ddc6-kube-api-access-kns6l\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.386557 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d505f96-7810-4021-9fd2-5abd46d6ddc6-secret-volume\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.386609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d505f96-7810-4021-9fd2-5abd46d6ddc6-config-volume\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.387690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d505f96-7810-4021-9fd2-5abd46d6ddc6-config-volume\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.393220 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d505f96-7810-4021-9fd2-5abd46d6ddc6-secret-volume\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.404133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kns6l\" (UniqueName: \"kubernetes.io/projected/8d505f96-7810-4021-9fd2-5abd46d6ddc6-kube-api-access-kns6l\") pod \"collect-profiles-29414310-mch6h\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.496610 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:00 crc kubenswrapper[4715]: I1204 14:30:00.972714 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h"] Dec 04 14:30:01 crc kubenswrapper[4715]: I1204 14:30:01.071047 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" event={"ID":"8d505f96-7810-4021-9fd2-5abd46d6ddc6","Type":"ContainerStarted","Data":"bb03cc58d6219b672ebb60ee6235749e56e8692d379d4291aceebb3625f6aa39"} Dec 04 14:30:02 crc kubenswrapper[4715]: I1204 14:30:02.081268 4715 generic.go:334] "Generic (PLEG): container finished" podID="8d505f96-7810-4021-9fd2-5abd46d6ddc6" containerID="0b98bc12264fb0cb65b8a4b9ada1c324327d832201e8e8ae163580aacb253eec" exitCode=0 Dec 04 14:30:02 crc kubenswrapper[4715]: I1204 14:30:02.081376 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" event={"ID":"8d505f96-7810-4021-9fd2-5abd46d6ddc6","Type":"ContainerDied","Data":"0b98bc12264fb0cb65b8a4b9ada1c324327d832201e8e8ae163580aacb253eec"} Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.440018 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.556965 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d505f96-7810-4021-9fd2-5abd46d6ddc6-secret-volume\") pod \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.557224 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kns6l\" (UniqueName: \"kubernetes.io/projected/8d505f96-7810-4021-9fd2-5abd46d6ddc6-kube-api-access-kns6l\") pod \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.557387 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d505f96-7810-4021-9fd2-5abd46d6ddc6-config-volume\") pod \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\" (UID: \"8d505f96-7810-4021-9fd2-5abd46d6ddc6\") " Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.558104 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d505f96-7810-4021-9fd2-5abd46d6ddc6-config-volume" (OuterVolumeSpecName: "config-volume") pod "8d505f96-7810-4021-9fd2-5abd46d6ddc6" (UID: "8d505f96-7810-4021-9fd2-5abd46d6ddc6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.569870 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d505f96-7810-4021-9fd2-5abd46d6ddc6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8d505f96-7810-4021-9fd2-5abd46d6ddc6" (UID: "8d505f96-7810-4021-9fd2-5abd46d6ddc6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.570337 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d505f96-7810-4021-9fd2-5abd46d6ddc6-kube-api-access-kns6l" (OuterVolumeSpecName: "kube-api-access-kns6l") pod "8d505f96-7810-4021-9fd2-5abd46d6ddc6" (UID: "8d505f96-7810-4021-9fd2-5abd46d6ddc6"). InnerVolumeSpecName "kube-api-access-kns6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.660142 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kns6l\" (UniqueName: \"kubernetes.io/projected/8d505f96-7810-4021-9fd2-5abd46d6ddc6-kube-api-access-kns6l\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.660198 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d505f96-7810-4021-9fd2-5abd46d6ddc6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:03 crc kubenswrapper[4715]: I1204 14:30:03.660208 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d505f96-7810-4021-9fd2-5abd46d6ddc6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:04 crc kubenswrapper[4715]: I1204 14:30:04.099659 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" event={"ID":"8d505f96-7810-4021-9fd2-5abd46d6ddc6","Type":"ContainerDied","Data":"bb03cc58d6219b672ebb60ee6235749e56e8692d379d4291aceebb3625f6aa39"} Dec 04 14:30:04 crc kubenswrapper[4715]: I1204 14:30:04.099709 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb03cc58d6219b672ebb60ee6235749e56e8692d379d4291aceebb3625f6aa39" Dec 04 14:30:04 crc kubenswrapper[4715]: I1204 14:30:04.099717 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h" Dec 04 14:30:07 crc kubenswrapper[4715]: I1204 14:30:07.180509 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:30:07 crc kubenswrapper[4715]: E1204 14:30:07.181352 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.048784 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wbgxz"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.058971 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-59d3-account-create-update-xm5ts"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.070848 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4x84z"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.079917 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-42a0-account-create-update-jhj8z"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.092180 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-ccnsf"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.103513 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-59d3-account-create-update-xm5ts"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.115013 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4x84z"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.123012 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-ccnsf"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.130446 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-42a0-account-create-update-jhj8z"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.142726 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-01e0-account-create-update-h84ff"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.152512 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wbgxz"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.161496 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-01e0-account-create-update-h84ff"] Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.201428 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13239d2e-d649-4a1c-aafc-fbdc135d7c2f" path="/var/lib/kubelet/pods/13239d2e-d649-4a1c-aafc-fbdc135d7c2f/volumes" Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.203062 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2139a8dc-19bb-453f-a293-02b7f5944810" path="/var/lib/kubelet/pods/2139a8dc-19bb-453f-a293-02b7f5944810/volumes" Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.204687 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584d122e-147a-48a3-993e-6ecd0ee9208a" path="/var/lib/kubelet/pods/584d122e-147a-48a3-993e-6ecd0ee9208a/volumes" Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.205641 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49248bc-1dc1-45c2-b0a5-6886331d3c54" path="/var/lib/kubelet/pods/d49248bc-1dc1-45c2-b0a5-6886331d3c54/volumes" Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.207323 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e7bb9b-d971-4641-b02d-0e17b0afe36f" path="/var/lib/kubelet/pods/f1e7bb9b-d971-4641-b02d-0e17b0afe36f/volumes" Dec 04 14:30:13 crc kubenswrapper[4715]: I1204 14:30:13.209054 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff544d83-9fa6-444c-bb1e-212318e8dc6a" path="/var/lib/kubelet/pods/ff544d83-9fa6-444c-bb1e-212318e8dc6a/volumes" Dec 04 14:30:21 crc kubenswrapper[4715]: I1204 14:30:21.180854 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:30:21 crc kubenswrapper[4715]: E1204 14:30:21.181598 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:30:21 crc kubenswrapper[4715]: I1204 14:30:21.246325 4715 generic.go:334] "Generic (PLEG): container finished" podID="7f4cab97-fa52-47c4-ad8e-7469556526cc" containerID="a78b88786d94fb93939e6c72721556388a06ae6b04b4706277fedbe2a6946c4d" exitCode=0 Dec 04 14:30:21 crc kubenswrapper[4715]: I1204 14:30:21.246376 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" event={"ID":"7f4cab97-fa52-47c4-ad8e-7469556526cc","Type":"ContainerDied","Data":"a78b88786d94fb93939e6c72721556388a06ae6b04b4706277fedbe2a6946c4d"} Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.693701 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.832367 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bfvk\" (UniqueName: \"kubernetes.io/projected/7f4cab97-fa52-47c4-ad8e-7469556526cc-kube-api-access-5bfvk\") pod \"7f4cab97-fa52-47c4-ad8e-7469556526cc\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.832638 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-ssh-key\") pod \"7f4cab97-fa52-47c4-ad8e-7469556526cc\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.832738 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-inventory\") pod \"7f4cab97-fa52-47c4-ad8e-7469556526cc\" (UID: \"7f4cab97-fa52-47c4-ad8e-7469556526cc\") " Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.840432 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4cab97-fa52-47c4-ad8e-7469556526cc-kube-api-access-5bfvk" (OuterVolumeSpecName: "kube-api-access-5bfvk") pod "7f4cab97-fa52-47c4-ad8e-7469556526cc" (UID: "7f4cab97-fa52-47c4-ad8e-7469556526cc"). InnerVolumeSpecName "kube-api-access-5bfvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.863751 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f4cab97-fa52-47c4-ad8e-7469556526cc" (UID: "7f4cab97-fa52-47c4-ad8e-7469556526cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.870926 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-inventory" (OuterVolumeSpecName: "inventory") pod "7f4cab97-fa52-47c4-ad8e-7469556526cc" (UID: "7f4cab97-fa52-47c4-ad8e-7469556526cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.936692 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.936754 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4cab97-fa52-47c4-ad8e-7469556526cc-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:22 crc kubenswrapper[4715]: I1204 14:30:22.936767 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bfvk\" (UniqueName: \"kubernetes.io/projected/7f4cab97-fa52-47c4-ad8e-7469556526cc-kube-api-access-5bfvk\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.268812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" event={"ID":"7f4cab97-fa52-47c4-ad8e-7469556526cc","Type":"ContainerDied","Data":"43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57"} Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.268928 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.268930 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6t47w" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.365117 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p"] Dec 04 14:30:23 crc kubenswrapper[4715]: E1204 14:30:23.365583 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4cab97-fa52-47c4-ad8e-7469556526cc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.365602 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4cab97-fa52-47c4-ad8e-7469556526cc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:23 crc kubenswrapper[4715]: E1204 14:30:23.365620 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d505f96-7810-4021-9fd2-5abd46d6ddc6" containerName="collect-profiles" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.365628 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d505f96-7810-4021-9fd2-5abd46d6ddc6" containerName="collect-profiles" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.365845 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4cab97-fa52-47c4-ad8e-7469556526cc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.365866 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d505f96-7810-4021-9fd2-5abd46d6ddc6" containerName="collect-profiles" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.366545 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.370113 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.370623 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.370700 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.370862 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.375620 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p"] Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.452946 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.453141 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmck2\" (UniqueName: \"kubernetes.io/projected/ee373f57-aa34-4c0a-86d4-5113c10984df-kube-api-access-gmck2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.453203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.556009 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.556493 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmck2\" (UniqueName: \"kubernetes.io/projected/ee373f57-aa34-4c0a-86d4-5113c10984df-kube-api-access-gmck2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.556565 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.563878 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.571656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.587962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmck2\" (UniqueName: \"kubernetes.io/projected/ee373f57-aa34-4c0a-86d4-5113c10984df-kube-api-access-gmck2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cq24p\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:23 crc kubenswrapper[4715]: I1204 14:30:23.692867 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:24 crc kubenswrapper[4715]: I1204 14:30:24.250654 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p"] Dec 04 14:30:24 crc kubenswrapper[4715]: I1204 14:30:24.258540 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:30:24 crc kubenswrapper[4715]: I1204 14:30:24.281187 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" event={"ID":"ee373f57-aa34-4c0a-86d4-5113c10984df","Type":"ContainerStarted","Data":"37b1c43bf636ebdc69868854a734bb002e7a902fc6ec4a434c62ed40833e217c"} Dec 04 14:30:25 crc kubenswrapper[4715]: I1204 14:30:25.293465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" event={"ID":"ee373f57-aa34-4c0a-86d4-5113c10984df","Type":"ContainerStarted","Data":"b98ad05c795c34fa39f0aaba483fce0f821e354fc01ba0d15804f8a77183dbbf"} Dec 04 14:30:25 crc kubenswrapper[4715]: I1204 14:30:25.315821 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" podStartSLOduration=1.900867255 podStartE2EDuration="2.315800349s" podCreationTimestamp="2025-12-04 14:30:23 +0000 UTC" firstStartedPulling="2025-12-04 14:30:24.258346179 +0000 UTC m=+1981.327064394" lastFinishedPulling="2025-12-04 14:30:24.673279273 +0000 UTC m=+1981.741997488" observedRunningTime="2025-12-04 14:30:25.312004197 +0000 UTC m=+1982.380722412" watchObservedRunningTime="2025-12-04 14:30:25.315800349 +0000 UTC m=+1982.384518564" Dec 04 14:30:30 crc kubenswrapper[4715]: E1204 14:30:30.225766 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice/crio-43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57\": RecentStats: unable to find data in memory cache]" Dec 04 14:30:30 crc kubenswrapper[4715]: I1204 14:30:30.352978 4715 generic.go:334] "Generic (PLEG): container finished" podID="ee373f57-aa34-4c0a-86d4-5113c10984df" containerID="b98ad05c795c34fa39f0aaba483fce0f821e354fc01ba0d15804f8a77183dbbf" exitCode=0 Dec 04 14:30:30 crc kubenswrapper[4715]: I1204 14:30:30.353140 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" event={"ID":"ee373f57-aa34-4c0a-86d4-5113c10984df","Type":"ContainerDied","Data":"b98ad05c795c34fa39f0aaba483fce0f821e354fc01ba0d15804f8a77183dbbf"} Dec 04 14:30:31 crc kubenswrapper[4715]: I1204 14:30:31.878839 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.020780 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-ssh-key\") pod \"ee373f57-aa34-4c0a-86d4-5113c10984df\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.021372 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmck2\" (UniqueName: \"kubernetes.io/projected/ee373f57-aa34-4c0a-86d4-5113c10984df-kube-api-access-gmck2\") pod \"ee373f57-aa34-4c0a-86d4-5113c10984df\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.021477 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-inventory\") pod \"ee373f57-aa34-4c0a-86d4-5113c10984df\" (UID: \"ee373f57-aa34-4c0a-86d4-5113c10984df\") " Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.027357 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee373f57-aa34-4c0a-86d4-5113c10984df-kube-api-access-gmck2" (OuterVolumeSpecName: "kube-api-access-gmck2") pod "ee373f57-aa34-4c0a-86d4-5113c10984df" (UID: "ee373f57-aa34-4c0a-86d4-5113c10984df"). InnerVolumeSpecName "kube-api-access-gmck2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.049566 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee373f57-aa34-4c0a-86d4-5113c10984df" (UID: "ee373f57-aa34-4c0a-86d4-5113c10984df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.050707 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-inventory" (OuterVolumeSpecName: "inventory") pod "ee373f57-aa34-4c0a-86d4-5113c10984df" (UID: "ee373f57-aa34-4c0a-86d4-5113c10984df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.124294 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmck2\" (UniqueName: \"kubernetes.io/projected/ee373f57-aa34-4c0a-86d4-5113c10984df-kube-api-access-gmck2\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.124330 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.124340 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee373f57-aa34-4c0a-86d4-5113c10984df-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.181147 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:30:32 crc kubenswrapper[4715]: E1204 14:30:32.181448 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.371308 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" event={"ID":"ee373f57-aa34-4c0a-86d4-5113c10984df","Type":"ContainerDied","Data":"37b1c43bf636ebdc69868854a734bb002e7a902fc6ec4a434c62ed40833e217c"} Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.371349 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b1c43bf636ebdc69868854a734bb002e7a902fc6ec4a434c62ed40833e217c" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.371416 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cq24p" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.451112 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js"] Dec 04 14:30:32 crc kubenswrapper[4715]: E1204 14:30:32.451594 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee373f57-aa34-4c0a-86d4-5113c10984df" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.451612 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee373f57-aa34-4c0a-86d4-5113c10984df" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.451860 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee373f57-aa34-4c0a-86d4-5113c10984df" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.452733 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.455668 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.455702 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.455892 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.457225 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.459629 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js"] Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.529975 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wqrg\" (UniqueName: \"kubernetes.io/projected/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-kube-api-access-9wqrg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.530102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.530158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.632293 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wqrg\" (UniqueName: \"kubernetes.io/projected/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-kube-api-access-9wqrg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.632399 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.632455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.636690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.637010 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.653940 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wqrg\" (UniqueName: \"kubernetes.io/projected/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-kube-api-access-9wqrg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6l4js\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:32 crc kubenswrapper[4715]: I1204 14:30:32.784062 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:30:33 crc kubenswrapper[4715]: I1204 14:30:33.376258 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js"] Dec 04 14:30:34 crc kubenswrapper[4715]: I1204 14:30:34.390612 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" event={"ID":"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2","Type":"ContainerStarted","Data":"5c7b3b70aeb0a0f79dac89df30281231a7a8cbd6dc1cab2a5e78151a91f56c0b"} Dec 04 14:30:34 crc kubenswrapper[4715]: I1204 14:30:34.391250 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" event={"ID":"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2","Type":"ContainerStarted","Data":"684b674fc18000b327412ddeec1c1b8f0ae4ab1bdd5e1bd02beb659c1347150e"} Dec 04 14:30:34 crc kubenswrapper[4715]: I1204 14:30:34.410347 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" podStartSLOduration=1.9393717719999999 podStartE2EDuration="2.410328406s" podCreationTimestamp="2025-12-04 14:30:32 +0000 UTC" firstStartedPulling="2025-12-04 14:30:33.380967402 +0000 UTC m=+1990.449685617" lastFinishedPulling="2025-12-04 14:30:33.851924036 +0000 UTC m=+1990.920642251" observedRunningTime="2025-12-04 14:30:34.408331602 +0000 UTC m=+1991.477049817" watchObservedRunningTime="2025-12-04 14:30:34.410328406 +0000 UTC m=+1991.479046621" Dec 04 14:30:38 crc kubenswrapper[4715]: I1204 14:30:38.417181 4715 scope.go:117] "RemoveContainer" containerID="d98000028662d7255e40983fd272c975c47a90a63b1956dca5731c2fc5ec5540" Dec 04 14:30:38 crc kubenswrapper[4715]: I1204 14:30:38.445402 4715 scope.go:117] "RemoveContainer" containerID="196ac5af78f8eec8cc7897c6da207f2c04d7f87ad6219743bfafb851eedbc235" Dec 04 14:30:38 crc kubenswrapper[4715]: I1204 14:30:38.496209 4715 scope.go:117] "RemoveContainer" containerID="bd271d86020cadb182cd20b55e5f1a3660bb60b7b70733a1ab48dd7fa5d04273" Dec 04 14:30:38 crc kubenswrapper[4715]: I1204 14:30:38.556555 4715 scope.go:117] "RemoveContainer" containerID="f1b186f73b402152779c2509ccc329465c4fd3c1d6c039a8d5ccfa24e80628c9" Dec 04 14:30:38 crc kubenswrapper[4715]: I1204 14:30:38.593014 4715 scope.go:117] "RemoveContainer" containerID="ba408ab6eeb9b2334ad7a5ffb2d27ea1e1b2d5007dd845cf518b54e9879253a0" Dec 04 14:30:38 crc kubenswrapper[4715]: I1204 14:30:38.635817 4715 scope.go:117] "RemoveContainer" containerID="673c9251cbe5e8bf0d54b5ba990a3a137dac691cd8675e5c827f151f4d8a2db9" Dec 04 14:30:40 crc kubenswrapper[4715]: E1204 14:30:40.497172 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice/crio-43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57\": RecentStats: unable to find data in memory cache]" Dec 04 14:30:46 crc kubenswrapper[4715]: I1204 14:30:46.180675 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:30:46 crc kubenswrapper[4715]: I1204 14:30:46.495729 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"e52fdf637b0ee5a2201c3751e1bf6e2b0117b3148977de7bf1774674b74cb9be"} Dec 04 14:30:47 crc kubenswrapper[4715]: I1204 14:30:47.046695 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fkv2x"] Dec 04 14:30:47 crc kubenswrapper[4715]: I1204 14:30:47.054635 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fkv2x"] Dec 04 14:30:47 crc kubenswrapper[4715]: I1204 14:30:47.194525 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d036e63f-d6a8-47cc-a19f-71a1dccf7698" path="/var/lib/kubelet/pods/d036e63f-d6a8-47cc-a19f-71a1dccf7698/volumes" Dec 04 14:30:50 crc kubenswrapper[4715]: E1204 14:30:50.775527 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice/crio-43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice\": RecentStats: unable to find data in memory cache]" Dec 04 14:31:01 crc kubenswrapper[4715]: E1204 14:31:01.051527 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice/crio-43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57\": RecentStats: unable to find data in memory cache]" Dec 04 14:31:10 crc kubenswrapper[4715]: I1204 14:31:10.704382 4715 generic.go:334] "Generic (PLEG): container finished" podID="b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" containerID="5c7b3b70aeb0a0f79dac89df30281231a7a8cbd6dc1cab2a5e78151a91f56c0b" exitCode=0 Dec 04 14:31:10 crc kubenswrapper[4715]: I1204 14:31:10.704911 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" event={"ID":"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2","Type":"ContainerDied","Data":"5c7b3b70aeb0a0f79dac89df30281231a7a8cbd6dc1cab2a5e78151a91f56c0b"} Dec 04 14:31:11 crc kubenswrapper[4715]: E1204 14:31:11.298142 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice/crio-43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice\": RecentStats: unable to find data in memory cache]" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.046085 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2h5kl"] Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.055861 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2h5kl"] Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.113679 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.196955 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-inventory\") pod \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.197224 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-ssh-key\") pod \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.197280 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wqrg\" (UniqueName: \"kubernetes.io/projected/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-kube-api-access-9wqrg\") pod \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\" (UID: \"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2\") " Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.203111 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-kube-api-access-9wqrg" (OuterVolumeSpecName: "kube-api-access-9wqrg") pod "b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" (UID: "b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2"). InnerVolumeSpecName "kube-api-access-9wqrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.227659 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-inventory" (OuterVolumeSpecName: "inventory") pod "b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" (UID: "b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.229823 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" (UID: "b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.299777 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.299806 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.299817 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wqrg\" (UniqueName: \"kubernetes.io/projected/b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2-kube-api-access-9wqrg\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.726148 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" event={"ID":"b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2","Type":"ContainerDied","Data":"684b674fc18000b327412ddeec1c1b8f0ae4ab1bdd5e1bd02beb659c1347150e"} Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.726195 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="684b674fc18000b327412ddeec1c1b8f0ae4ab1bdd5e1bd02beb659c1347150e" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.726297 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6l4js" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.826249 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq"] Dec 04 14:31:12 crc kubenswrapper[4715]: E1204 14:31:12.826973 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.826990 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.827215 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.827976 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.832146 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.832197 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.832332 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.836627 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.838126 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq"] Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.913721 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.913890 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:12 crc kubenswrapper[4715]: I1204 14:31:12.913929 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8tmk\" (UniqueName: \"kubernetes.io/projected/337aa155-3f3f-4e84-a350-29d27ef13260-kube-api-access-x8tmk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.017208 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.017296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.017322 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8tmk\" (UniqueName: \"kubernetes.io/projected/337aa155-3f3f-4e84-a350-29d27ef13260-kube-api-access-x8tmk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.021327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.026776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.035867 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8tmk\" (UniqueName: \"kubernetes.io/projected/337aa155-3f3f-4e84-a350-29d27ef13260-kube-api-access-x8tmk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.154217 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.192553 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54688981-4eb1-49e5-ac13-39f4f5429bab" path="/var/lib/kubelet/pods/54688981-4eb1-49e5-ac13-39f4f5429bab/volumes" Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.659440 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq"] Dec 04 14:31:13 crc kubenswrapper[4715]: I1204 14:31:13.734290 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" event={"ID":"337aa155-3f3f-4e84-a350-29d27ef13260","Type":"ContainerStarted","Data":"d01b8569d4f8c20c26901680d128a80afc514c3a47d66e7141bc4bd29e3f799c"} Dec 04 14:31:14 crc kubenswrapper[4715]: I1204 14:31:14.744423 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" event={"ID":"337aa155-3f3f-4e84-a350-29d27ef13260","Type":"ContainerStarted","Data":"e84259124e13ced312e92a132dac215130e3cd2b625d8ef0c8b8fc240ef96f93"} Dec 04 14:31:14 crc kubenswrapper[4715]: I1204 14:31:14.766783 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" podStartSLOduration=2.284537888 podStartE2EDuration="2.766765935s" podCreationTimestamp="2025-12-04 14:31:12 +0000 UTC" firstStartedPulling="2025-12-04 14:31:13.666540931 +0000 UTC m=+2030.735259146" lastFinishedPulling="2025-12-04 14:31:14.148768978 +0000 UTC m=+2031.217487193" observedRunningTime="2025-12-04 14:31:14.760370365 +0000 UTC m=+2031.829088590" watchObservedRunningTime="2025-12-04 14:31:14.766765935 +0000 UTC m=+2031.835484150" Dec 04 14:31:18 crc kubenswrapper[4715]: I1204 14:31:18.031775 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6pd76"] Dec 04 14:31:18 crc kubenswrapper[4715]: I1204 14:31:18.040752 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6pd76"] Dec 04 14:31:19 crc kubenswrapper[4715]: I1204 14:31:19.191304 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95dcc828-2041-48f7-9f22-31911926f820" path="/var/lib/kubelet/pods/95dcc828-2041-48f7-9f22-31911926f820/volumes" Dec 04 14:31:21 crc kubenswrapper[4715]: E1204 14:31:21.544479 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4cab97_fa52_47c4_ad8e_7469556526cc.slice/crio-43180bc5af7ab52ad868fb014f17b979a0aed056d3bfd7bebb19aab474445f57\": RecentStats: unable to find data in memory cache]" Dec 04 14:31:38 crc kubenswrapper[4715]: I1204 14:31:38.783114 4715 scope.go:117] "RemoveContainer" containerID="d0f96d0cb13b050343af85e299f56a398e2ec0a27b0c24accd18dd9f10609ed1" Dec 04 14:31:38 crc kubenswrapper[4715]: I1204 14:31:38.829810 4715 scope.go:117] "RemoveContainer" containerID="79dc9fd13a77abbf6d0dfb63977787e93fa92bbacd9c3463ee04b10efbc4beda" Dec 04 14:31:38 crc kubenswrapper[4715]: I1204 14:31:38.881795 4715 scope.go:117] "RemoveContainer" containerID="d6657b9b09615afca6ca840be9dbe4c674110c0a7d955559eed125246a562fd8" Dec 04 14:31:56 crc kubenswrapper[4715]: I1204 14:31:56.044934 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-7hlzp"] Dec 04 14:31:56 crc kubenswrapper[4715]: I1204 14:31:56.053152 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-7hlzp"] Dec 04 14:31:57 crc kubenswrapper[4715]: I1204 14:31:57.214328 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e3b168-dd77-4ba1-bdd4-dd4072b30dfb" path="/var/lib/kubelet/pods/87e3b168-dd77-4ba1-bdd4-dd4072b30dfb/volumes" Dec 04 14:32:02 crc kubenswrapper[4715]: I1204 14:32:02.168706 4715 generic.go:334] "Generic (PLEG): container finished" podID="337aa155-3f3f-4e84-a350-29d27ef13260" containerID="e84259124e13ced312e92a132dac215130e3cd2b625d8ef0c8b8fc240ef96f93" exitCode=0 Dec 04 14:32:02 crc kubenswrapper[4715]: I1204 14:32:02.168826 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" event={"ID":"337aa155-3f3f-4e84-a350-29d27ef13260","Type":"ContainerDied","Data":"e84259124e13ced312e92a132dac215130e3cd2b625d8ef0c8b8fc240ef96f93"} Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.604342 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.799069 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-ssh-key\") pod \"337aa155-3f3f-4e84-a350-29d27ef13260\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.799221 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-inventory\") pod \"337aa155-3f3f-4e84-a350-29d27ef13260\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.799379 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8tmk\" (UniqueName: \"kubernetes.io/projected/337aa155-3f3f-4e84-a350-29d27ef13260-kube-api-access-x8tmk\") pod \"337aa155-3f3f-4e84-a350-29d27ef13260\" (UID: \"337aa155-3f3f-4e84-a350-29d27ef13260\") " Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.805508 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337aa155-3f3f-4e84-a350-29d27ef13260-kube-api-access-x8tmk" (OuterVolumeSpecName: "kube-api-access-x8tmk") pod "337aa155-3f3f-4e84-a350-29d27ef13260" (UID: "337aa155-3f3f-4e84-a350-29d27ef13260"). InnerVolumeSpecName "kube-api-access-x8tmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.829532 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-inventory" (OuterVolumeSpecName: "inventory") pod "337aa155-3f3f-4e84-a350-29d27ef13260" (UID: "337aa155-3f3f-4e84-a350-29d27ef13260"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.829901 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "337aa155-3f3f-4e84-a350-29d27ef13260" (UID: "337aa155-3f3f-4e84-a350-29d27ef13260"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.901873 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.901916 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/337aa155-3f3f-4e84-a350-29d27ef13260-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:03 crc kubenswrapper[4715]: I1204 14:32:03.901930 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8tmk\" (UniqueName: \"kubernetes.io/projected/337aa155-3f3f-4e84-a350-29d27ef13260-kube-api-access-x8tmk\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.185252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" event={"ID":"337aa155-3f3f-4e84-a350-29d27ef13260","Type":"ContainerDied","Data":"d01b8569d4f8c20c26901680d128a80afc514c3a47d66e7141bc4bd29e3f799c"} Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.185284 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.185289 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d01b8569d4f8c20c26901680d128a80afc514c3a47d66e7141bc4bd29e3f799c" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.281431 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4wlm4"] Dec 04 14:32:04 crc kubenswrapper[4715]: E1204 14:32:04.282185 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337aa155-3f3f-4e84-a350-29d27ef13260" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.282279 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="337aa155-3f3f-4e84-a350-29d27ef13260" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.282724 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="337aa155-3f3f-4e84-a350-29d27ef13260" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.283684 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.287394 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.287565 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.287844 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.288109 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.293276 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4wlm4"] Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.313055 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.313220 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.313282 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jndc\" (UniqueName: \"kubernetes.io/projected/5058ed17-7d86-4b90-a41b-668b6407054a-kube-api-access-6jndc\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.415541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.415677 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.415726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jndc\" (UniqueName: \"kubernetes.io/projected/5058ed17-7d86-4b90-a41b-668b6407054a-kube-api-access-6jndc\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.421474 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.432870 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jndc\" (UniqueName: \"kubernetes.io/projected/5058ed17-7d86-4b90-a41b-668b6407054a-kube-api-access-6jndc\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.446148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4wlm4\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:04 crc kubenswrapper[4715]: I1204 14:32:04.605733 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:05 crc kubenswrapper[4715]: I1204 14:32:05.201284 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4wlm4"] Dec 04 14:32:06 crc kubenswrapper[4715]: I1204 14:32:06.206106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" event={"ID":"5058ed17-7d86-4b90-a41b-668b6407054a","Type":"ContainerStarted","Data":"4e42a918f1861b8e10d7183f2af77c6cbe2e1f9c7333c42ce237ca1279c5b068"} Dec 04 14:32:06 crc kubenswrapper[4715]: I1204 14:32:06.206433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" event={"ID":"5058ed17-7d86-4b90-a41b-668b6407054a","Type":"ContainerStarted","Data":"533bd4ff22cb408a1ef950cefd23d2d9565e04da840ab798072652cc432ab5f7"} Dec 04 14:32:06 crc kubenswrapper[4715]: I1204 14:32:06.231086 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" podStartSLOduration=1.802312733 podStartE2EDuration="2.231069214s" podCreationTimestamp="2025-12-04 14:32:04 +0000 UTC" firstStartedPulling="2025-12-04 14:32:05.209107222 +0000 UTC m=+2082.277825437" lastFinishedPulling="2025-12-04 14:32:05.637863703 +0000 UTC m=+2082.706581918" observedRunningTime="2025-12-04 14:32:06.222389699 +0000 UTC m=+2083.291107914" watchObservedRunningTime="2025-12-04 14:32:06.231069214 +0000 UTC m=+2083.299787429" Dec 04 14:32:12 crc kubenswrapper[4715]: E1204 14:32:12.844532 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5058ed17_7d86_4b90_a41b_668b6407054a.slice/crio-4e42a918f1861b8e10d7183f2af77c6cbe2e1f9c7333c42ce237ca1279c5b068.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5058ed17_7d86_4b90_a41b_668b6407054a.slice/crio-conmon-4e42a918f1861b8e10d7183f2af77c6cbe2e1f9c7333c42ce237ca1279c5b068.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:32:13 crc kubenswrapper[4715]: I1204 14:32:13.275155 4715 generic.go:334] "Generic (PLEG): container finished" podID="5058ed17-7d86-4b90-a41b-668b6407054a" containerID="4e42a918f1861b8e10d7183f2af77c6cbe2e1f9c7333c42ce237ca1279c5b068" exitCode=0 Dec 04 14:32:13 crc kubenswrapper[4715]: I1204 14:32:13.275206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" event={"ID":"5058ed17-7d86-4b90-a41b-668b6407054a","Type":"ContainerDied","Data":"4e42a918f1861b8e10d7183f2af77c6cbe2e1f9c7333c42ce237ca1279c5b068"} Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.731769 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.853829 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jndc\" (UniqueName: \"kubernetes.io/projected/5058ed17-7d86-4b90-a41b-668b6407054a-kube-api-access-6jndc\") pod \"5058ed17-7d86-4b90-a41b-668b6407054a\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.854067 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-ssh-key-openstack-edpm-ipam\") pod \"5058ed17-7d86-4b90-a41b-668b6407054a\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.854092 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-inventory-0\") pod \"5058ed17-7d86-4b90-a41b-668b6407054a\" (UID: \"5058ed17-7d86-4b90-a41b-668b6407054a\") " Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.859567 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5058ed17-7d86-4b90-a41b-668b6407054a-kube-api-access-6jndc" (OuterVolumeSpecName: "kube-api-access-6jndc") pod "5058ed17-7d86-4b90-a41b-668b6407054a" (UID: "5058ed17-7d86-4b90-a41b-668b6407054a"). InnerVolumeSpecName "kube-api-access-6jndc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.883287 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5058ed17-7d86-4b90-a41b-668b6407054a" (UID: "5058ed17-7d86-4b90-a41b-668b6407054a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.887392 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5058ed17-7d86-4b90-a41b-668b6407054a" (UID: "5058ed17-7d86-4b90-a41b-668b6407054a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.956785 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jndc\" (UniqueName: \"kubernetes.io/projected/5058ed17-7d86-4b90-a41b-668b6407054a-kube-api-access-6jndc\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.956824 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:14 crc kubenswrapper[4715]: I1204 14:32:14.956835 4715 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5058ed17-7d86-4b90-a41b-668b6407054a-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.293900 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" event={"ID":"5058ed17-7d86-4b90-a41b-668b6407054a","Type":"ContainerDied","Data":"533bd4ff22cb408a1ef950cefd23d2d9565e04da840ab798072652cc432ab5f7"} Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.293943 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="533bd4ff22cb408a1ef950cefd23d2d9565e04da840ab798072652cc432ab5f7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.294331 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4wlm4" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.368451 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7"] Dec 04 14:32:15 crc kubenswrapper[4715]: E1204 14:32:15.369335 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5058ed17-7d86-4b90-a41b-668b6407054a" containerName="ssh-known-hosts-edpm-deployment" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.369425 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5058ed17-7d86-4b90-a41b-668b6407054a" containerName="ssh-known-hosts-edpm-deployment" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.369710 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5058ed17-7d86-4b90-a41b-668b6407054a" containerName="ssh-known-hosts-edpm-deployment" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.370642 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.372537 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.372551 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.372978 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.373412 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.378575 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7"] Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.467703 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.468013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snn5d\" (UniqueName: \"kubernetes.io/projected/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-kube-api-access-snn5d\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.468221 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.570384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.570790 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.570922 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snn5d\" (UniqueName: \"kubernetes.io/projected/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-kube-api-access-snn5d\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.576192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.576529 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.591195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snn5d\" (UniqueName: \"kubernetes.io/projected/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-kube-api-access-snn5d\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fpzs7\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:15 crc kubenswrapper[4715]: I1204 14:32:15.687408 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:16 crc kubenswrapper[4715]: I1204 14:32:16.253122 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7"] Dec 04 14:32:16 crc kubenswrapper[4715]: I1204 14:32:16.311698 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" event={"ID":"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1","Type":"ContainerStarted","Data":"ac4e470fadfa5d72d89cd1cacfac407bf491e3d4877fd86027dec098123f5d67"} Dec 04 14:32:17 crc kubenswrapper[4715]: I1204 14:32:17.323950 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" event={"ID":"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1","Type":"ContainerStarted","Data":"04f30f21558a9186343e5ffb25011436be8bbb0458021a5612f1023416471fad"} Dec 04 14:32:17 crc kubenswrapper[4715]: I1204 14:32:17.345999 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" podStartSLOduration=1.932423729 podStartE2EDuration="2.345978322s" podCreationTimestamp="2025-12-04 14:32:15 +0000 UTC" firstStartedPulling="2025-12-04 14:32:16.27533602 +0000 UTC m=+2093.344054235" lastFinishedPulling="2025-12-04 14:32:16.688890623 +0000 UTC m=+2093.757608828" observedRunningTime="2025-12-04 14:32:17.343681707 +0000 UTC m=+2094.412399922" watchObservedRunningTime="2025-12-04 14:32:17.345978322 +0000 UTC m=+2094.414696537" Dec 04 14:32:25 crc kubenswrapper[4715]: I1204 14:32:25.396238 4715 generic.go:334] "Generic (PLEG): container finished" podID="d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" containerID="04f30f21558a9186343e5ffb25011436be8bbb0458021a5612f1023416471fad" exitCode=0 Dec 04 14:32:25 crc kubenswrapper[4715]: I1204 14:32:25.396332 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" event={"ID":"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1","Type":"ContainerDied","Data":"04f30f21558a9186343e5ffb25011436be8bbb0458021a5612f1023416471fad"} Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.808643 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.921612 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snn5d\" (UniqueName: \"kubernetes.io/projected/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-kube-api-access-snn5d\") pod \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.921764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-inventory\") pod \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.921806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-ssh-key\") pod \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\" (UID: \"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1\") " Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.927576 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-kube-api-access-snn5d" (OuterVolumeSpecName: "kube-api-access-snn5d") pod "d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" (UID: "d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1"). InnerVolumeSpecName "kube-api-access-snn5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.955200 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-inventory" (OuterVolumeSpecName: "inventory") pod "d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" (UID: "d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:26 crc kubenswrapper[4715]: I1204 14:32:26.957656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" (UID: "d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.026662 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snn5d\" (UniqueName: \"kubernetes.io/projected/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-kube-api-access-snn5d\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.026769 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.026786 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.416262 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" event={"ID":"d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1","Type":"ContainerDied","Data":"ac4e470fadfa5d72d89cd1cacfac407bf491e3d4877fd86027dec098123f5d67"} Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.416309 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac4e470fadfa5d72d89cd1cacfac407bf491e3d4877fd86027dec098123f5d67" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.416328 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fpzs7" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.493539 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj"] Dec 04 14:32:27 crc kubenswrapper[4715]: E1204 14:32:27.494080 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.494105 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.494354 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.495296 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.497479 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.497762 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.498019 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.498511 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.502412 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj"] Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.537100 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.537309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.537440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7bg8\" (UniqueName: \"kubernetes.io/projected/baed77e4-c68b-4663-b4c2-0f0de712d963-kube-api-access-s7bg8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.638957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.639069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7bg8\" (UniqueName: \"kubernetes.io/projected/baed77e4-c68b-4663-b4c2-0f0de712d963-kube-api-access-s7bg8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.639135 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.643258 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.643444 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.656495 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7bg8\" (UniqueName: \"kubernetes.io/projected/baed77e4-c68b-4663-b4c2-0f0de712d963-kube-api-access-s7bg8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:27 crc kubenswrapper[4715]: I1204 14:32:27.814713 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:28 crc kubenswrapper[4715]: I1204 14:32:28.335191 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj"] Dec 04 14:32:28 crc kubenswrapper[4715]: I1204 14:32:28.430790 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" event={"ID":"baed77e4-c68b-4663-b4c2-0f0de712d963","Type":"ContainerStarted","Data":"709356b6b5e4f0bdedbc742a091e971123e27415c4fcce9df6f04252d7f81de8"} Dec 04 14:32:29 crc kubenswrapper[4715]: I1204 14:32:29.440812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" event={"ID":"baed77e4-c68b-4663-b4c2-0f0de712d963","Type":"ContainerStarted","Data":"bcb72e78c718228fcd218db80cb228bb84946e1adef00e36afd0f665343410d6"} Dec 04 14:32:29 crc kubenswrapper[4715]: I1204 14:32:29.467770 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" podStartSLOduration=2.031744494 podStartE2EDuration="2.467752699s" podCreationTimestamp="2025-12-04 14:32:27 +0000 UTC" firstStartedPulling="2025-12-04 14:32:28.343299261 +0000 UTC m=+2105.412017476" lastFinishedPulling="2025-12-04 14:32:28.779307466 +0000 UTC m=+2105.848025681" observedRunningTime="2025-12-04 14:32:29.457580152 +0000 UTC m=+2106.526298367" watchObservedRunningTime="2025-12-04 14:32:29.467752699 +0000 UTC m=+2106.536470914" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.516387 4715 generic.go:334] "Generic (PLEG): container finished" podID="baed77e4-c68b-4663-b4c2-0f0de712d963" containerID="bcb72e78c718228fcd218db80cb228bb84946e1adef00e36afd0f665343410d6" exitCode=0 Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.516448 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" event={"ID":"baed77e4-c68b-4663-b4c2-0f0de712d963","Type":"ContainerDied","Data":"bcb72e78c718228fcd218db80cb228bb84946e1adef00e36afd0f665343410d6"} Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.609950 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7xzk"] Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.618219 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.633123 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7xzk"] Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.671148 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-catalog-content\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.671482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tztb\" (UniqueName: \"kubernetes.io/projected/335d39d8-0335-472e-b4e9-9a576ee884f8-kube-api-access-2tztb\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.671626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-utilities\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.773405 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-utilities\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.773588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-catalog-content\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.773652 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tztb\" (UniqueName: \"kubernetes.io/projected/335d39d8-0335-472e-b4e9-9a576ee884f8-kube-api-access-2tztb\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.774329 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-catalog-content\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.774408 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-utilities\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.814209 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tztb\" (UniqueName: \"kubernetes.io/projected/335d39d8-0335-472e-b4e9-9a576ee884f8-kube-api-access-2tztb\") pod \"redhat-operators-x7xzk\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:38 crc kubenswrapper[4715]: I1204 14:32:38.945551 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:39 crc kubenswrapper[4715]: I1204 14:32:39.020686 4715 scope.go:117] "RemoveContainer" containerID="e624bc6b5d84e244abb219db0871ef091414caa93421ce28f0ce179a22af2a24" Dec 04 14:32:39 crc kubenswrapper[4715]: I1204 14:32:39.431072 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7xzk"] Dec 04 14:32:39 crc kubenswrapper[4715]: I1204 14:32:39.526054 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerStarted","Data":"04b17696c04936a861c84902d0ff3bddf72d147b79f42fce84d6324d01b55151"} Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.026796 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.106696 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-inventory\") pod \"baed77e4-c68b-4663-b4c2-0f0de712d963\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.106818 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-ssh-key\") pod \"baed77e4-c68b-4663-b4c2-0f0de712d963\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.106876 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7bg8\" (UniqueName: \"kubernetes.io/projected/baed77e4-c68b-4663-b4c2-0f0de712d963-kube-api-access-s7bg8\") pod \"baed77e4-c68b-4663-b4c2-0f0de712d963\" (UID: \"baed77e4-c68b-4663-b4c2-0f0de712d963\") " Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.112651 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baed77e4-c68b-4663-b4c2-0f0de712d963-kube-api-access-s7bg8" (OuterVolumeSpecName: "kube-api-access-s7bg8") pod "baed77e4-c68b-4663-b4c2-0f0de712d963" (UID: "baed77e4-c68b-4663-b4c2-0f0de712d963"). InnerVolumeSpecName "kube-api-access-s7bg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.136982 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "baed77e4-c68b-4663-b4c2-0f0de712d963" (UID: "baed77e4-c68b-4663-b4c2-0f0de712d963"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.142992 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-inventory" (OuterVolumeSpecName: "inventory") pod "baed77e4-c68b-4663-b4c2-0f0de712d963" (UID: "baed77e4-c68b-4663-b4c2-0f0de712d963"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.211879 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.212215 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/baed77e4-c68b-4663-b4c2-0f0de712d963-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.212226 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7bg8\" (UniqueName: \"kubernetes.io/projected/baed77e4-c68b-4663-b4c2-0f0de712d963-kube-api-access-s7bg8\") on node \"crc\" DevicePath \"\"" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.540009 4715 generic.go:334] "Generic (PLEG): container finished" podID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerID="67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020" exitCode=0 Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.540109 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerDied","Data":"67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020"} Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.548114 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" event={"ID":"baed77e4-c68b-4663-b4c2-0f0de712d963","Type":"ContainerDied","Data":"709356b6b5e4f0bdedbc742a091e971123e27415c4fcce9df6f04252d7f81de8"} Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.548155 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.548161 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="709356b6b5e4f0bdedbc742a091e971123e27415c4fcce9df6f04252d7f81de8" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.710554 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv"] Dec 04 14:32:40 crc kubenswrapper[4715]: E1204 14:32:40.711112 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baed77e4-c68b-4663-b4c2-0f0de712d963" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.711140 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="baed77e4-c68b-4663-b4c2-0f0de712d963" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.711413 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="baed77e4-c68b-4663-b4c2-0f0de712d963" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.712258 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.718016 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.718173 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.718429 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.718482 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.718842 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.718961 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.719057 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.719425 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.734233 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv"] Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823576 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823624 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823667 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823742 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823776 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823803 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823819 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823849 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823919 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.823986 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fnrk\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-kube-api-access-7fnrk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.925836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.925897 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.925933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.925969 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.925992 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926056 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926109 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926135 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926235 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fnrk\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-kube-api-access-7fnrk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.926402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.932407 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.932889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.933466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.933922 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.934252 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.934337 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.934488 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.935153 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.935614 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.936697 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.939740 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.950641 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fnrk\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-kube-api-access-7fnrk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.952804 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:40 crc kubenswrapper[4715]: I1204 14:32:40.953414 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:41 crc kubenswrapper[4715]: I1204 14:32:41.040192 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:32:41 crc kubenswrapper[4715]: I1204 14:32:41.568572 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerStarted","Data":"84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac"} Dec 04 14:32:41 crc kubenswrapper[4715]: I1204 14:32:41.676929 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv"] Dec 04 14:32:41 crc kubenswrapper[4715]: W1204 14:32:41.704196 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd394ffb7_2675_4316_8388_cee93c9a37bc.slice/crio-cd8d8c68caae64ffebba0465f3e14a7b85aea1c59ee704e84cd2316ee3581037 WatchSource:0}: Error finding container cd8d8c68caae64ffebba0465f3e14a7b85aea1c59ee704e84cd2316ee3581037: Status 404 returned error can't find the container with id cd8d8c68caae64ffebba0465f3e14a7b85aea1c59ee704e84cd2316ee3581037 Dec 04 14:32:42 crc kubenswrapper[4715]: I1204 14:32:42.578964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" event={"ID":"d394ffb7-2675-4316-8388-cee93c9a37bc","Type":"ContainerStarted","Data":"cd8d8c68caae64ffebba0465f3e14a7b85aea1c59ee704e84cd2316ee3581037"} Dec 04 14:32:43 crc kubenswrapper[4715]: I1204 14:32:43.589381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" event={"ID":"d394ffb7-2675-4316-8388-cee93c9a37bc","Type":"ContainerStarted","Data":"4e7e14492d6fa7dd8c17edd3924c414dcdb090b2a6d3c8c19da7eabbfbe71479"} Dec 04 14:32:43 crc kubenswrapper[4715]: I1204 14:32:43.594091 4715 generic.go:334] "Generic (PLEG): container finished" podID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerID="84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac" exitCode=0 Dec 04 14:32:43 crc kubenswrapper[4715]: I1204 14:32:43.594154 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerDied","Data":"84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac"} Dec 04 14:32:43 crc kubenswrapper[4715]: I1204 14:32:43.612981 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" podStartSLOduration=2.9886641149999997 podStartE2EDuration="3.612955231s" podCreationTimestamp="2025-12-04 14:32:40 +0000 UTC" firstStartedPulling="2025-12-04 14:32:41.706710134 +0000 UTC m=+2118.775428349" lastFinishedPulling="2025-12-04 14:32:42.33100125 +0000 UTC m=+2119.399719465" observedRunningTime="2025-12-04 14:32:43.609403721 +0000 UTC m=+2120.678121956" watchObservedRunningTime="2025-12-04 14:32:43.612955231 +0000 UTC m=+2120.681673446" Dec 04 14:32:46 crc kubenswrapper[4715]: I1204 14:32:46.635400 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerStarted","Data":"20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018"} Dec 04 14:32:46 crc kubenswrapper[4715]: I1204 14:32:46.659624 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7xzk" podStartSLOduration=3.72248894 podStartE2EDuration="8.659600754s" podCreationTimestamp="2025-12-04 14:32:38 +0000 UTC" firstStartedPulling="2025-12-04 14:32:40.542775227 +0000 UTC m=+2117.611493442" lastFinishedPulling="2025-12-04 14:32:45.479887041 +0000 UTC m=+2122.548605256" observedRunningTime="2025-12-04 14:32:46.653563334 +0000 UTC m=+2123.722281549" watchObservedRunningTime="2025-12-04 14:32:46.659600754 +0000 UTC m=+2123.728318969" Dec 04 14:32:48 crc kubenswrapper[4715]: I1204 14:32:48.945838 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:48 crc kubenswrapper[4715]: I1204 14:32:48.946206 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:49 crc kubenswrapper[4715]: I1204 14:32:49.994102 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x7xzk" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="registry-server" probeResult="failure" output=< Dec 04 14:32:49 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 04 14:32:49 crc kubenswrapper[4715]: > Dec 04 14:32:58 crc kubenswrapper[4715]: I1204 14:32:58.995929 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:59 crc kubenswrapper[4715]: I1204 14:32:59.045352 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:32:59 crc kubenswrapper[4715]: I1204 14:32:59.234132 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7xzk"] Dec 04 14:33:00 crc kubenswrapper[4715]: I1204 14:33:00.746255 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7xzk" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="registry-server" containerID="cri-o://20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018" gracePeriod=2 Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.206633 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.382400 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tztb\" (UniqueName: \"kubernetes.io/projected/335d39d8-0335-472e-b4e9-9a576ee884f8-kube-api-access-2tztb\") pod \"335d39d8-0335-472e-b4e9-9a576ee884f8\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.382546 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-utilities\") pod \"335d39d8-0335-472e-b4e9-9a576ee884f8\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.382784 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-catalog-content\") pod \"335d39d8-0335-472e-b4e9-9a576ee884f8\" (UID: \"335d39d8-0335-472e-b4e9-9a576ee884f8\") " Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.383350 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-utilities" (OuterVolumeSpecName: "utilities") pod "335d39d8-0335-472e-b4e9-9a576ee884f8" (UID: "335d39d8-0335-472e-b4e9-9a576ee884f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.390214 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/335d39d8-0335-472e-b4e9-9a576ee884f8-kube-api-access-2tztb" (OuterVolumeSpecName: "kube-api-access-2tztb") pod "335d39d8-0335-472e-b4e9-9a576ee884f8" (UID: "335d39d8-0335-472e-b4e9-9a576ee884f8"). InnerVolumeSpecName "kube-api-access-2tztb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.485471 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tztb\" (UniqueName: \"kubernetes.io/projected/335d39d8-0335-472e-b4e9-9a576ee884f8-kube-api-access-2tztb\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.485520 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.493341 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "335d39d8-0335-472e-b4e9-9a576ee884f8" (UID: "335d39d8-0335-472e-b4e9-9a576ee884f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.587370 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335d39d8-0335-472e-b4e9-9a576ee884f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.757078 4715 generic.go:334] "Generic (PLEG): container finished" podID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerID="20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018" exitCode=0 Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.757123 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xzk" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.757157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerDied","Data":"20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018"} Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.757251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xzk" event={"ID":"335d39d8-0335-472e-b4e9-9a576ee884f8","Type":"ContainerDied","Data":"04b17696c04936a861c84902d0ff3bddf72d147b79f42fce84d6324d01b55151"} Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.757277 4715 scope.go:117] "RemoveContainer" containerID="20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.776989 4715 scope.go:117] "RemoveContainer" containerID="84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.790775 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7xzk"] Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.798884 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7xzk"] Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.816139 4715 scope.go:117] "RemoveContainer" containerID="67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.853640 4715 scope.go:117] "RemoveContainer" containerID="20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018" Dec 04 14:33:01 crc kubenswrapper[4715]: E1204 14:33:01.854275 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018\": container with ID starting with 20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018 not found: ID does not exist" containerID="20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.854410 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018"} err="failed to get container status \"20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018\": rpc error: code = NotFound desc = could not find container \"20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018\": container with ID starting with 20220c69927f3e1bcb0d494e5fa8c955425645a17b0128d1300d0e31f4975018 not found: ID does not exist" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.854499 4715 scope.go:117] "RemoveContainer" containerID="84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac" Dec 04 14:33:01 crc kubenswrapper[4715]: E1204 14:33:01.855345 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac\": container with ID starting with 84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac not found: ID does not exist" containerID="84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.855393 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac"} err="failed to get container status \"84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac\": rpc error: code = NotFound desc = could not find container \"84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac\": container with ID starting with 84543eeb8ee870b44d4bf02852586c17fc918e15de5257135fcc596b9b9be1ac not found: ID does not exist" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.855420 4715 scope.go:117] "RemoveContainer" containerID="67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020" Dec 04 14:33:01 crc kubenswrapper[4715]: E1204 14:33:01.855783 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020\": container with ID starting with 67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020 not found: ID does not exist" containerID="67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020" Dec 04 14:33:01 crc kubenswrapper[4715]: I1204 14:33:01.855806 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020"} err="failed to get container status \"67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020\": rpc error: code = NotFound desc = could not find container \"67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020\": container with ID starting with 67ac917d70d78fa4d517df871f0168d88fcc9de8045630b65909a9f178bdb020 not found: ID does not exist" Dec 04 14:33:03 crc kubenswrapper[4715]: I1204 14:33:03.190724 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" path="/var/lib/kubelet/pods/335d39d8-0335-472e-b4e9-9a576ee884f8/volumes" Dec 04 14:33:08 crc kubenswrapper[4715]: I1204 14:33:08.758370 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:33:08 crc kubenswrapper[4715]: I1204 14:33:08.759462 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:33:18 crc kubenswrapper[4715]: I1204 14:33:18.903264 4715 generic.go:334] "Generic (PLEG): container finished" podID="d394ffb7-2675-4316-8388-cee93c9a37bc" containerID="4e7e14492d6fa7dd8c17edd3924c414dcdb090b2a6d3c8c19da7eabbfbe71479" exitCode=0 Dec 04 14:33:18 crc kubenswrapper[4715]: I1204 14:33:18.903728 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" event={"ID":"d394ffb7-2675-4316-8388-cee93c9a37bc","Type":"ContainerDied","Data":"4e7e14492d6fa7dd8c17edd3924c414dcdb090b2a6d3c8c19da7eabbfbe71479"} Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.331576 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.456441 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-libvirt-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.456536 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.456619 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-inventory\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.456687 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-repo-setup-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.456729 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.456976 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-telemetry-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.457728 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fnrk\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-kube-api-access-7fnrk\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.457776 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-nova-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.457822 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-neutron-metadata-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.457899 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ssh-key\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.457924 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.457959 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ovn-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.458008 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.458055 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-bootstrap-combined-ca-bundle\") pod \"d394ffb7-2675-4316-8388-cee93c9a37bc\" (UID: \"d394ffb7-2675-4316-8388-cee93c9a37bc\") " Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.465497 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.465536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-kube-api-access-7fnrk" (OuterVolumeSpecName: "kube-api-access-7fnrk") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "kube-api-access-7fnrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.465465 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.465692 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.465738 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.465772 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.466587 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.467665 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.468366 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.468400 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.468784 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.469352 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.491694 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-inventory" (OuterVolumeSpecName: "inventory") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.492151 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d394ffb7-2675-4316-8388-cee93c9a37bc" (UID: "d394ffb7-2675-4316-8388-cee93c9a37bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560177 4715 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560211 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560223 4715 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560233 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fnrk\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-kube-api-access-7fnrk\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560241 4715 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560254 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560265 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560275 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560287 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560296 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560305 4715 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560313 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560322 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d394ffb7-2675-4316-8388-cee93c9a37bc-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.560330 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d394ffb7-2675-4316-8388-cee93c9a37bc-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.923911 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" event={"ID":"d394ffb7-2675-4316-8388-cee93c9a37bc","Type":"ContainerDied","Data":"cd8d8c68caae64ffebba0465f3e14a7b85aea1c59ee704e84cd2316ee3581037"} Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.923960 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd8d8c68caae64ffebba0465f3e14a7b85aea1c59ee704e84cd2316ee3581037" Dec 04 14:33:20 crc kubenswrapper[4715]: I1204 14:33:20.923978 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.022823 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62"] Dec 04 14:33:21 crc kubenswrapper[4715]: E1204 14:33:21.023379 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="extract-utilities" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.023404 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="extract-utilities" Dec 04 14:33:21 crc kubenswrapper[4715]: E1204 14:33:21.023424 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="extract-content" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.023433 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="extract-content" Dec 04 14:33:21 crc kubenswrapper[4715]: E1204 14:33:21.023478 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d394ffb7-2675-4316-8388-cee93c9a37bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.023488 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d394ffb7-2675-4316-8388-cee93c9a37bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 14:33:21 crc kubenswrapper[4715]: E1204 14:33:21.023504 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="registry-server" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.023511 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="registry-server" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.023736 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="335d39d8-0335-472e-b4e9-9a576ee884f8" containerName="registry-server" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.023779 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d394ffb7-2675-4316-8388-cee93c9a37bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.024610 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.026933 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.027426 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.027470 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.027587 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.029184 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.041131 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62"] Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.171448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.171503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9ntd\" (UniqueName: \"kubernetes.io/projected/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-kube-api-access-q9ntd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.171610 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.171655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.171730 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.273936 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.273994 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9ntd\" (UniqueName: \"kubernetes.io/projected/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-kube-api-access-q9ntd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.274070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.274090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.274146 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.275450 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.278046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.278274 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.279767 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.297190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9ntd\" (UniqueName: \"kubernetes.io/projected/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-kube-api-access-q9ntd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6qf62\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.342140 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.840117 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62"] Dec 04 14:33:21 crc kubenswrapper[4715]: I1204 14:33:21.933942 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" event={"ID":"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5","Type":"ContainerStarted","Data":"c50d68354c4909af939f6a0ec15af68d5a8bb2be8976a2e3a60c0c37a20c042b"} Dec 04 14:33:22 crc kubenswrapper[4715]: I1204 14:33:22.943079 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" event={"ID":"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5","Type":"ContainerStarted","Data":"825b00a02dcb238fb4948b1a76151c85a4c6cb8bba6edce195f502227b3e5c15"} Dec 04 14:33:22 crc kubenswrapper[4715]: I1204 14:33:22.973906 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" podStartSLOduration=1.548006985 podStartE2EDuration="1.973884764s" podCreationTimestamp="2025-12-04 14:33:21 +0000 UTC" firstStartedPulling="2025-12-04 14:33:21.852547665 +0000 UTC m=+2158.921265880" lastFinishedPulling="2025-12-04 14:33:22.278425444 +0000 UTC m=+2159.347143659" observedRunningTime="2025-12-04 14:33:22.958537022 +0000 UTC m=+2160.027255257" watchObservedRunningTime="2025-12-04 14:33:22.973884764 +0000 UTC m=+2160.042602989" Dec 04 14:33:38 crc kubenswrapper[4715]: I1204 14:33:38.758418 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:33:38 crc kubenswrapper[4715]: I1204 14:33:38.758872 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:34:08 crc kubenswrapper[4715]: I1204 14:34:08.758401 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:34:08 crc kubenswrapper[4715]: I1204 14:34:08.758936 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:34:08 crc kubenswrapper[4715]: I1204 14:34:08.758989 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:34:08 crc kubenswrapper[4715]: I1204 14:34:08.759788 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e52fdf637b0ee5a2201c3751e1bf6e2b0117b3148977de7bf1774674b74cb9be"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:34:08 crc kubenswrapper[4715]: I1204 14:34:08.759850 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://e52fdf637b0ee5a2201c3751e1bf6e2b0117b3148977de7bf1774674b74cb9be" gracePeriod=600 Dec 04 14:34:09 crc kubenswrapper[4715]: I1204 14:34:09.399709 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="e52fdf637b0ee5a2201c3751e1bf6e2b0117b3148977de7bf1774674b74cb9be" exitCode=0 Dec 04 14:34:09 crc kubenswrapper[4715]: I1204 14:34:09.399781 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"e52fdf637b0ee5a2201c3751e1bf6e2b0117b3148977de7bf1774674b74cb9be"} Dec 04 14:34:09 crc kubenswrapper[4715]: I1204 14:34:09.400066 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a"} Dec 04 14:34:09 crc kubenswrapper[4715]: I1204 14:34:09.400089 4715 scope.go:117] "RemoveContainer" containerID="57ff10a31286d9cc8bdad182b61e3ae420e5fa338f0d4f929500a9b88cba0d6b" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.423164 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wlj9m"] Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.427133 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.444153 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wlj9m"] Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.531075 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" event={"ID":"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5","Type":"ContainerDied","Data":"825b00a02dcb238fb4948b1a76151c85a4c6cb8bba6edce195f502227b3e5c15"} Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.531087 4715 generic.go:334] "Generic (PLEG): container finished" podID="d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" containerID="825b00a02dcb238fb4948b1a76151c85a4c6cb8bba6edce195f502227b3e5c15" exitCode=0 Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.594348 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-utilities\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.594591 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9sv8\" (UniqueName: \"kubernetes.io/projected/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-kube-api-access-q9sv8\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.594697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-catalog-content\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.697011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-utilities\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.697492 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9sv8\" (UniqueName: \"kubernetes.io/projected/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-kube-api-access-q9sv8\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.697542 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-utilities\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.698149 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-catalog-content\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.699155 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-catalog-content\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.720740 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9sv8\" (UniqueName: \"kubernetes.io/projected/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-kube-api-access-q9sv8\") pod \"community-operators-wlj9m\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:23 crc kubenswrapper[4715]: I1204 14:34:23.764541 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:24 crc kubenswrapper[4715]: I1204 14:34:24.355581 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wlj9m"] Dec 04 14:34:24 crc kubenswrapper[4715]: I1204 14:34:24.540884 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerStarted","Data":"7b42b21aea569e19c409bae3969bf4f5c1c5e052fc6314e7f1427608d451e10f"} Dec 04 14:34:24 crc kubenswrapper[4715]: I1204 14:34:24.982517 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.126514 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9ntd\" (UniqueName: \"kubernetes.io/projected/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-kube-api-access-q9ntd\") pod \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.126839 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovn-combined-ca-bundle\") pod \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.127106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovncontroller-config-0\") pod \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.127267 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-inventory\") pod \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.127382 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ssh-key\") pod \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\" (UID: \"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5\") " Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.139906 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" (UID: "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.139964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-kube-api-access-q9ntd" (OuterVolumeSpecName: "kube-api-access-q9ntd") pod "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" (UID: "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5"). InnerVolumeSpecName "kube-api-access-q9ntd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.154450 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" (UID: "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.159063 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-inventory" (OuterVolumeSpecName: "inventory") pod "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" (UID: "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.160227 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" (UID: "d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.230364 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9ntd\" (UniqueName: \"kubernetes.io/projected/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-kube-api-access-q9ntd\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.230395 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.230411 4715 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.230424 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.230436 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.549372 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" event={"ID":"d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5","Type":"ContainerDied","Data":"c50d68354c4909af939f6a0ec15af68d5a8bb2be8976a2e3a60c0c37a20c042b"} Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.549413 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c50d68354c4909af939f6a0ec15af68d5a8bb2be8976a2e3a60c0c37a20c042b" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.549436 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6qf62" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.551085 4715 generic.go:334] "Generic (PLEG): container finished" podID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerID="a9886c892999ebefe8a112d2bf554a29a92d33078f01994f78a781eef4b2ae9e" exitCode=0 Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.551127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerDied","Data":"a9886c892999ebefe8a112d2bf554a29a92d33078f01994f78a781eef4b2ae9e"} Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.639698 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl"] Dec 04 14:34:25 crc kubenswrapper[4715]: E1204 14:34:25.640679 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.640789 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.641142 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.641995 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.647981 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.649003 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.649149 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.650364 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.650523 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.650672 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.658364 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl"] Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.743187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.743314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.743339 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.743455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs9lc\" (UniqueName: \"kubernetes.io/projected/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-kube-api-access-xs9lc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.743492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.743537 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.844849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.844948 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.845025 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.845077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.845185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs9lc\" (UniqueName: \"kubernetes.io/projected/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-kube-api-access-xs9lc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.845226 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.851009 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.851068 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.851410 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.852802 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.858660 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.861991 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs9lc\" (UniqueName: \"kubernetes.io/projected/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-kube-api-access-xs9lc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:25 crc kubenswrapper[4715]: I1204 14:34:25.964716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:34:26 crc kubenswrapper[4715]: I1204 14:34:26.518979 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl"] Dec 04 14:34:26 crc kubenswrapper[4715]: W1204 14:34:26.520021 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8a7fc4f_4f90_4023_98af_a9613dbce1cb.slice/crio-a4e01937bbff73d36f52e2a76f9d947d2d09a674fb9a7ee304371a9f69cce0c3 WatchSource:0}: Error finding container a4e01937bbff73d36f52e2a76f9d947d2d09a674fb9a7ee304371a9f69cce0c3: Status 404 returned error can't find the container with id a4e01937bbff73d36f52e2a76f9d947d2d09a674fb9a7ee304371a9f69cce0c3 Dec 04 14:34:26 crc kubenswrapper[4715]: I1204 14:34:26.564902 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerStarted","Data":"803c1fbd18a8131c90011ded7c72d39019fd95782555e15fe4870512934e5196"} Dec 04 14:34:26 crc kubenswrapper[4715]: I1204 14:34:26.566890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" event={"ID":"e8a7fc4f-4f90-4023-98af-a9613dbce1cb","Type":"ContainerStarted","Data":"a4e01937bbff73d36f52e2a76f9d947d2d09a674fb9a7ee304371a9f69cce0c3"} Dec 04 14:34:27 crc kubenswrapper[4715]: I1204 14:34:27.577007 4715 generic.go:334] "Generic (PLEG): container finished" podID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerID="803c1fbd18a8131c90011ded7c72d39019fd95782555e15fe4870512934e5196" exitCode=0 Dec 04 14:34:27 crc kubenswrapper[4715]: I1204 14:34:27.577071 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerDied","Data":"803c1fbd18a8131c90011ded7c72d39019fd95782555e15fe4870512934e5196"} Dec 04 14:34:27 crc kubenswrapper[4715]: I1204 14:34:27.581007 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" event={"ID":"e8a7fc4f-4f90-4023-98af-a9613dbce1cb","Type":"ContainerStarted","Data":"db3cc7e3246a1b817cedad56e482ea3a5856901e23cb901206d8375c35d9a966"} Dec 04 14:34:27 crc kubenswrapper[4715]: I1204 14:34:27.629527 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" podStartSLOduration=2.092749227 podStartE2EDuration="2.629508789s" podCreationTimestamp="2025-12-04 14:34:25 +0000 UTC" firstStartedPulling="2025-12-04 14:34:26.523320236 +0000 UTC m=+2223.592038451" lastFinishedPulling="2025-12-04 14:34:27.060079778 +0000 UTC m=+2224.128798013" observedRunningTime="2025-12-04 14:34:27.622519742 +0000 UTC m=+2224.691237967" watchObservedRunningTime="2025-12-04 14:34:27.629508789 +0000 UTC m=+2224.698227004" Dec 04 14:34:28 crc kubenswrapper[4715]: I1204 14:34:28.592330 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerStarted","Data":"ba303bcf87a5fa5904272869adc57b2963d551d8ef9d75911c6b3905ba988b72"} Dec 04 14:34:28 crc kubenswrapper[4715]: I1204 14:34:28.619850 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wlj9m" podStartSLOduration=3.027043924 podStartE2EDuration="5.619824279s" podCreationTimestamp="2025-12-04 14:34:23 +0000 UTC" firstStartedPulling="2025-12-04 14:34:25.553240245 +0000 UTC m=+2222.621958460" lastFinishedPulling="2025-12-04 14:34:28.1460206 +0000 UTC m=+2225.214738815" observedRunningTime="2025-12-04 14:34:28.611064742 +0000 UTC m=+2225.679782957" watchObservedRunningTime="2025-12-04 14:34:28.619824279 +0000 UTC m=+2225.688542494" Dec 04 14:34:33 crc kubenswrapper[4715]: I1204 14:34:33.765288 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:33 crc kubenswrapper[4715]: I1204 14:34:33.766623 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:33 crc kubenswrapper[4715]: I1204 14:34:33.831330 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:34 crc kubenswrapper[4715]: I1204 14:34:34.687434 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:34 crc kubenswrapper[4715]: I1204 14:34:34.728593 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wlj9m"] Dec 04 14:34:36 crc kubenswrapper[4715]: I1204 14:34:36.661484 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wlj9m" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="registry-server" containerID="cri-o://ba303bcf87a5fa5904272869adc57b2963d551d8ef9d75911c6b3905ba988b72" gracePeriod=2 Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.670443 4715 generic.go:334] "Generic (PLEG): container finished" podID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerID="ba303bcf87a5fa5904272869adc57b2963d551d8ef9d75911c6b3905ba988b72" exitCode=0 Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.671023 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerDied","Data":"ba303bcf87a5fa5904272869adc57b2963d551d8ef9d75911c6b3905ba988b72"} Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.671064 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wlj9m" event={"ID":"92c9c18a-6d2a-4dce-aea2-b32c47555aa0","Type":"ContainerDied","Data":"7b42b21aea569e19c409bae3969bf4f5c1c5e052fc6314e7f1427608d451e10f"} Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.671076 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b42b21aea569e19c409bae3969bf4f5c1c5e052fc6314e7f1427608d451e10f" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.681867 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.801506 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-catalog-content\") pod \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.801547 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-utilities\") pod \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.801860 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9sv8\" (UniqueName: \"kubernetes.io/projected/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-kube-api-access-q9sv8\") pod \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\" (UID: \"92c9c18a-6d2a-4dce-aea2-b32c47555aa0\") " Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.818917 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-kube-api-access-q9sv8" (OuterVolumeSpecName: "kube-api-access-q9sv8") pod "92c9c18a-6d2a-4dce-aea2-b32c47555aa0" (UID: "92c9c18a-6d2a-4dce-aea2-b32c47555aa0"). InnerVolumeSpecName "kube-api-access-q9sv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.819682 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-utilities" (OuterVolumeSpecName: "utilities") pod "92c9c18a-6d2a-4dce-aea2-b32c47555aa0" (UID: "92c9c18a-6d2a-4dce-aea2-b32c47555aa0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.889966 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92c9c18a-6d2a-4dce-aea2-b32c47555aa0" (UID: "92c9c18a-6d2a-4dce-aea2-b32c47555aa0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.911149 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9sv8\" (UniqueName: \"kubernetes.io/projected/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-kube-api-access-q9sv8\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.911185 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:37 crc kubenswrapper[4715]: I1204 14:34:37.911198 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c9c18a-6d2a-4dce-aea2-b32c47555aa0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:38 crc kubenswrapper[4715]: I1204 14:34:38.677845 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wlj9m" Dec 04 14:34:38 crc kubenswrapper[4715]: I1204 14:34:38.709709 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wlj9m"] Dec 04 14:34:38 crc kubenswrapper[4715]: I1204 14:34:38.720522 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wlj9m"] Dec 04 14:34:39 crc kubenswrapper[4715]: I1204 14:34:39.193601 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" path="/var/lib/kubelet/pods/92c9c18a-6d2a-4dce-aea2-b32c47555aa0/volumes" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.606892 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gbrs9"] Dec 04 14:34:45 crc kubenswrapper[4715]: E1204 14:34:45.607885 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="registry-server" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.607902 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="registry-server" Dec 04 14:34:45 crc kubenswrapper[4715]: E1204 14:34:45.607924 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="extract-content" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.607930 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="extract-content" Dec 04 14:34:45 crc kubenswrapper[4715]: E1204 14:34:45.607951 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="extract-utilities" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.607960 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="extract-utilities" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.608209 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c9c18a-6d2a-4dce-aea2-b32c47555aa0" containerName="registry-server" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.609867 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.618829 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gbrs9"] Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.772388 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-utilities\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.772564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-catalog-content\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.772630 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9kg8\" (UniqueName: \"kubernetes.io/projected/502120c0-d257-449f-b82b-aaf4e8fb2408-kube-api-access-x9kg8\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.874069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-catalog-content\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.874177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9kg8\" (UniqueName: \"kubernetes.io/projected/502120c0-d257-449f-b82b-aaf4e8fb2408-kube-api-access-x9kg8\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.874251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-utilities\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.874719 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-catalog-content\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.874820 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-utilities\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.903995 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9kg8\" (UniqueName: \"kubernetes.io/projected/502120c0-d257-449f-b82b-aaf4e8fb2408-kube-api-access-x9kg8\") pod \"redhat-marketplace-gbrs9\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:45 crc kubenswrapper[4715]: I1204 14:34:45.937074 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:46 crc kubenswrapper[4715]: I1204 14:34:46.442742 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gbrs9"] Dec 04 14:34:46 crc kubenswrapper[4715]: I1204 14:34:46.768455 4715 generic.go:334] "Generic (PLEG): container finished" podID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerID="b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197" exitCode=0 Dec 04 14:34:46 crc kubenswrapper[4715]: I1204 14:34:46.768544 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gbrs9" event={"ID":"502120c0-d257-449f-b82b-aaf4e8fb2408","Type":"ContainerDied","Data":"b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197"} Dec 04 14:34:46 crc kubenswrapper[4715]: I1204 14:34:46.768616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gbrs9" event={"ID":"502120c0-d257-449f-b82b-aaf4e8fb2408","Type":"ContainerStarted","Data":"0704d7acbed975486be88a363999b52edb4702bcddff42190cf7e121cb9f400b"} Dec 04 14:34:47 crc kubenswrapper[4715]: I1204 14:34:47.785065 4715 generic.go:334] "Generic (PLEG): container finished" podID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerID="4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c" exitCode=0 Dec 04 14:34:47 crc kubenswrapper[4715]: I1204 14:34:47.785183 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gbrs9" event={"ID":"502120c0-d257-449f-b82b-aaf4e8fb2408","Type":"ContainerDied","Data":"4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c"} Dec 04 14:34:48 crc kubenswrapper[4715]: I1204 14:34:48.796303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gbrs9" event={"ID":"502120c0-d257-449f-b82b-aaf4e8fb2408","Type":"ContainerStarted","Data":"369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104"} Dec 04 14:34:48 crc kubenswrapper[4715]: I1204 14:34:48.821761 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gbrs9" podStartSLOduration=2.414439814 podStartE2EDuration="3.821738834s" podCreationTimestamp="2025-12-04 14:34:45 +0000 UTC" firstStartedPulling="2025-12-04 14:34:46.773597912 +0000 UTC m=+2243.842316127" lastFinishedPulling="2025-12-04 14:34:48.180896932 +0000 UTC m=+2245.249615147" observedRunningTime="2025-12-04 14:34:48.8169662 +0000 UTC m=+2245.885684425" watchObservedRunningTime="2025-12-04 14:34:48.821738834 +0000 UTC m=+2245.890457069" Dec 04 14:34:55 crc kubenswrapper[4715]: I1204 14:34:55.937495 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:55 crc kubenswrapper[4715]: I1204 14:34:55.938000 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:55 crc kubenswrapper[4715]: I1204 14:34:55.985450 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:56 crc kubenswrapper[4715]: I1204 14:34:56.944719 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:57 crc kubenswrapper[4715]: I1204 14:34:57.004676 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gbrs9"] Dec 04 14:34:58 crc kubenswrapper[4715]: I1204 14:34:58.904344 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gbrs9" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="registry-server" containerID="cri-o://369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104" gracePeriod=2 Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.405314 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.547419 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9kg8\" (UniqueName: \"kubernetes.io/projected/502120c0-d257-449f-b82b-aaf4e8fb2408-kube-api-access-x9kg8\") pod \"502120c0-d257-449f-b82b-aaf4e8fb2408\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.547633 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-catalog-content\") pod \"502120c0-d257-449f-b82b-aaf4e8fb2408\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.547671 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-utilities\") pod \"502120c0-d257-449f-b82b-aaf4e8fb2408\" (UID: \"502120c0-d257-449f-b82b-aaf4e8fb2408\") " Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.548791 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-utilities" (OuterVolumeSpecName: "utilities") pod "502120c0-d257-449f-b82b-aaf4e8fb2408" (UID: "502120c0-d257-449f-b82b-aaf4e8fb2408"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.557329 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502120c0-d257-449f-b82b-aaf4e8fb2408-kube-api-access-x9kg8" (OuterVolumeSpecName: "kube-api-access-x9kg8") pod "502120c0-d257-449f-b82b-aaf4e8fb2408" (UID: "502120c0-d257-449f-b82b-aaf4e8fb2408"). InnerVolumeSpecName "kube-api-access-x9kg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.569716 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "502120c0-d257-449f-b82b-aaf4e8fb2408" (UID: "502120c0-d257-449f-b82b-aaf4e8fb2408"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.649803 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.649842 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502120c0-d257-449f-b82b-aaf4e8fb2408-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.649853 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9kg8\" (UniqueName: \"kubernetes.io/projected/502120c0-d257-449f-b82b-aaf4e8fb2408-kube-api-access-x9kg8\") on node \"crc\" DevicePath \"\"" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.916681 4715 generic.go:334] "Generic (PLEG): container finished" podID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerID="369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104" exitCode=0 Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.916730 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gbrs9" event={"ID":"502120c0-d257-449f-b82b-aaf4e8fb2408","Type":"ContainerDied","Data":"369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104"} Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.916787 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gbrs9" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.916811 4715 scope.go:117] "RemoveContainer" containerID="369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.916799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gbrs9" event={"ID":"502120c0-d257-449f-b82b-aaf4e8fb2408","Type":"ContainerDied","Data":"0704d7acbed975486be88a363999b52edb4702bcddff42190cf7e121cb9f400b"} Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.956908 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gbrs9"] Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.963497 4715 scope.go:117] "RemoveContainer" containerID="4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c" Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.968080 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gbrs9"] Dec 04 14:34:59 crc kubenswrapper[4715]: I1204 14:34:59.982634 4715 scope.go:117] "RemoveContainer" containerID="b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197" Dec 04 14:35:00 crc kubenswrapper[4715]: I1204 14:35:00.033517 4715 scope.go:117] "RemoveContainer" containerID="369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104" Dec 04 14:35:00 crc kubenswrapper[4715]: E1204 14:35:00.034133 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104\": container with ID starting with 369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104 not found: ID does not exist" containerID="369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104" Dec 04 14:35:00 crc kubenswrapper[4715]: I1204 14:35:00.034272 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104"} err="failed to get container status \"369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104\": rpc error: code = NotFound desc = could not find container \"369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104\": container with ID starting with 369cb20f4d89b74c4ab3a50008177e5ae8451573990050d7652a35b4689e1104 not found: ID does not exist" Dec 04 14:35:00 crc kubenswrapper[4715]: I1204 14:35:00.034424 4715 scope.go:117] "RemoveContainer" containerID="4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c" Dec 04 14:35:00 crc kubenswrapper[4715]: E1204 14:35:00.034987 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c\": container with ID starting with 4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c not found: ID does not exist" containerID="4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c" Dec 04 14:35:00 crc kubenswrapper[4715]: I1204 14:35:00.035023 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c"} err="failed to get container status \"4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c\": rpc error: code = NotFound desc = could not find container \"4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c\": container with ID starting with 4eb7d13f440b084469f4ead88af3f6ed3789c8a68178f90eb2e5f8af46c63d2c not found: ID does not exist" Dec 04 14:35:00 crc kubenswrapper[4715]: I1204 14:35:00.035065 4715 scope.go:117] "RemoveContainer" containerID="b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197" Dec 04 14:35:00 crc kubenswrapper[4715]: E1204 14:35:00.035384 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197\": container with ID starting with b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197 not found: ID does not exist" containerID="b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197" Dec 04 14:35:00 crc kubenswrapper[4715]: I1204 14:35:00.035494 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197"} err="failed to get container status \"b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197\": rpc error: code = NotFound desc = could not find container \"b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197\": container with ID starting with b8a2516918b8ab49fd356fa8d32e268c4ca826642e5bbb73f5536ca2985f1197 not found: ID does not exist" Dec 04 14:35:01 crc kubenswrapper[4715]: I1204 14:35:01.193703 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" path="/var/lib/kubelet/pods/502120c0-d257-449f-b82b-aaf4e8fb2408/volumes" Dec 04 14:35:15 crc kubenswrapper[4715]: I1204 14:35:15.084862 4715 generic.go:334] "Generic (PLEG): container finished" podID="e8a7fc4f-4f90-4023-98af-a9613dbce1cb" containerID="db3cc7e3246a1b817cedad56e482ea3a5856901e23cb901206d8375c35d9a966" exitCode=0 Dec 04 14:35:15 crc kubenswrapper[4715]: I1204 14:35:15.084968 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" event={"ID":"e8a7fc4f-4f90-4023-98af-a9613dbce1cb","Type":"ContainerDied","Data":"db3cc7e3246a1b817cedad56e482ea3a5856901e23cb901206d8375c35d9a966"} Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.552841 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.606963 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.607981 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-ssh-key\") pod \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.608072 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-nova-metadata-neutron-config-0\") pod \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.608262 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs9lc\" (UniqueName: \"kubernetes.io/projected/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-kube-api-access-xs9lc\") pod \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.608294 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-metadata-combined-ca-bundle\") pod \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.608368 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-inventory\") pod \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\" (UID: \"e8a7fc4f-4f90-4023-98af-a9613dbce1cb\") " Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.626226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-kube-api-access-xs9lc" (OuterVolumeSpecName: "kube-api-access-xs9lc") pod "e8a7fc4f-4f90-4023-98af-a9613dbce1cb" (UID: "e8a7fc4f-4f90-4023-98af-a9613dbce1cb"). InnerVolumeSpecName "kube-api-access-xs9lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.630216 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e8a7fc4f-4f90-4023-98af-a9613dbce1cb" (UID: "e8a7fc4f-4f90-4023-98af-a9613dbce1cb"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.636395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-inventory" (OuterVolumeSpecName: "inventory") pod "e8a7fc4f-4f90-4023-98af-a9613dbce1cb" (UID: "e8a7fc4f-4f90-4023-98af-a9613dbce1cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.640445 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e8a7fc4f-4f90-4023-98af-a9613dbce1cb" (UID: "e8a7fc4f-4f90-4023-98af-a9613dbce1cb"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.642318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e8a7fc4f-4f90-4023-98af-a9613dbce1cb" (UID: "e8a7fc4f-4f90-4023-98af-a9613dbce1cb"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.648557 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e8a7fc4f-4f90-4023-98af-a9613dbce1cb" (UID: "e8a7fc4f-4f90-4023-98af-a9613dbce1cb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.711712 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.711985 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs9lc\" (UniqueName: \"kubernetes.io/projected/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-kube-api-access-xs9lc\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.711995 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.712006 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.712016 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:16 crc kubenswrapper[4715]: I1204 14:35:16.712027 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8a7fc4f-4f90-4023-98af-a9613dbce1cb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.111136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" event={"ID":"e8a7fc4f-4f90-4023-98af-a9613dbce1cb","Type":"ContainerDied","Data":"a4e01937bbff73d36f52e2a76f9d947d2d09a674fb9a7ee304371a9f69cce0c3"} Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.111223 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e01937bbff73d36f52e2a76f9d947d2d09a674fb9a7ee304371a9f69cce0c3" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.111316 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.244519 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt"] Dec 04 14:35:17 crc kubenswrapper[4715]: E1204 14:35:17.245024 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a7fc4f-4f90-4023-98af-a9613dbce1cb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.245085 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a7fc4f-4f90-4023-98af-a9613dbce1cb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 14:35:17 crc kubenswrapper[4715]: E1204 14:35:17.245112 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="registry-server" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.245125 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="registry-server" Dec 04 14:35:17 crc kubenswrapper[4715]: E1204 14:35:17.245165 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="extract-utilities" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.245177 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="extract-utilities" Dec 04 14:35:17 crc kubenswrapper[4715]: E1204 14:35:17.245200 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="extract-content" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.245211 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="extract-content" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.245531 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8a7fc4f-4f90-4023-98af-a9613dbce1cb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.245577 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="502120c0-d257-449f-b82b-aaf4e8fb2408" containerName="registry-server" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.246600 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.248946 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.250397 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.250404 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.250572 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.250655 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.259577 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt"] Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.325272 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjksl\" (UniqueName: \"kubernetes.io/projected/4a13c152-6be0-4261-ae2f-3b7b3458005a-kube-api-access-hjksl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.325335 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.325440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.325502 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.325549 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.426862 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjksl\" (UniqueName: \"kubernetes.io/projected/4a13c152-6be0-4261-ae2f-3b7b3458005a-kube-api-access-hjksl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.426939 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.427000 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.427081 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.427112 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.432956 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.433078 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.443451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.445948 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.448481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjksl\" (UniqueName: \"kubernetes.io/projected/4a13c152-6be0-4261-ae2f-3b7b3458005a-kube-api-access-hjksl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:17 crc kubenswrapper[4715]: I1204 14:35:17.574932 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:35:18 crc kubenswrapper[4715]: I1204 14:35:18.198688 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt"] Dec 04 14:35:19 crc kubenswrapper[4715]: I1204 14:35:19.132103 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" event={"ID":"4a13c152-6be0-4261-ae2f-3b7b3458005a","Type":"ContainerStarted","Data":"624bcaaadd2f8fdaaeb88af1dc6b8e84c2ec3ea67fe06d924297e2cacf0737f7"} Dec 04 14:35:19 crc kubenswrapper[4715]: I1204 14:35:19.132404 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" event={"ID":"4a13c152-6be0-4261-ae2f-3b7b3458005a","Type":"ContainerStarted","Data":"7f35719d53fac254e6ed9aba8e32ca096f577eb1cb79a19f49524a613ef0acdb"} Dec 04 14:35:19 crc kubenswrapper[4715]: I1204 14:35:19.150990 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" podStartSLOduration=1.745044931 podStartE2EDuration="2.150968622s" podCreationTimestamp="2025-12-04 14:35:17 +0000 UTC" firstStartedPulling="2025-12-04 14:35:18.198432738 +0000 UTC m=+2275.267150953" lastFinishedPulling="2025-12-04 14:35:18.604356429 +0000 UTC m=+2275.673074644" observedRunningTime="2025-12-04 14:35:19.147391005 +0000 UTC m=+2276.216109220" watchObservedRunningTime="2025-12-04 14:35:19.150968622 +0000 UTC m=+2276.219686837" Dec 04 14:35:37 crc kubenswrapper[4715]: I1204 14:35:37.881817 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qptfm"] Dec 04 14:35:37 crc kubenswrapper[4715]: I1204 14:35:37.884141 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:37 crc kubenswrapper[4715]: I1204 14:35:37.911945 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qptfm"] Dec 04 14:35:37 crc kubenswrapper[4715]: I1204 14:35:37.984304 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd4nl\" (UniqueName: \"kubernetes.io/projected/c8244f07-d253-45fc-94af-09de9d42b8ce-kube-api-access-jd4nl\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:37 crc kubenswrapper[4715]: I1204 14:35:37.984507 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-catalog-content\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:37 crc kubenswrapper[4715]: I1204 14:35:37.984551 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-utilities\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.086453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-catalog-content\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.086756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-utilities\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.086898 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd4nl\" (UniqueName: \"kubernetes.io/projected/c8244f07-d253-45fc-94af-09de9d42b8ce-kube-api-access-jd4nl\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.087107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-catalog-content\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.087261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-utilities\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.124935 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd4nl\" (UniqueName: \"kubernetes.io/projected/c8244f07-d253-45fc-94af-09de9d42b8ce-kube-api-access-jd4nl\") pod \"certified-operators-qptfm\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.208109 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:38 crc kubenswrapper[4715]: I1204 14:35:38.796588 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qptfm"] Dec 04 14:35:38 crc kubenswrapper[4715]: W1204 14:35:38.797189 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8244f07_d253_45fc_94af_09de9d42b8ce.slice/crio-a35ccf830daa1fc95bcc6d1dd86cc84c3a8e19f4c1dee25564fdf9b742e3ba07 WatchSource:0}: Error finding container a35ccf830daa1fc95bcc6d1dd86cc84c3a8e19f4c1dee25564fdf9b742e3ba07: Status 404 returned error can't find the container with id a35ccf830daa1fc95bcc6d1dd86cc84c3a8e19f4c1dee25564fdf9b742e3ba07 Dec 04 14:35:39 crc kubenswrapper[4715]: I1204 14:35:39.353890 4715 generic.go:334] "Generic (PLEG): container finished" podID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerID="8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16" exitCode=0 Dec 04 14:35:39 crc kubenswrapper[4715]: I1204 14:35:39.353938 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerDied","Data":"8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16"} Dec 04 14:35:39 crc kubenswrapper[4715]: I1204 14:35:39.354212 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerStarted","Data":"a35ccf830daa1fc95bcc6d1dd86cc84c3a8e19f4c1dee25564fdf9b742e3ba07"} Dec 04 14:35:39 crc kubenswrapper[4715]: I1204 14:35:39.356466 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:35:40 crc kubenswrapper[4715]: I1204 14:35:40.372434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerStarted","Data":"082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1"} Dec 04 14:35:41 crc kubenswrapper[4715]: I1204 14:35:41.386089 4715 generic.go:334] "Generic (PLEG): container finished" podID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerID="082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1" exitCode=0 Dec 04 14:35:41 crc kubenswrapper[4715]: I1204 14:35:41.386166 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerDied","Data":"082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1"} Dec 04 14:35:42 crc kubenswrapper[4715]: I1204 14:35:42.397316 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerStarted","Data":"a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095"} Dec 04 14:35:42 crc kubenswrapper[4715]: I1204 14:35:42.433957 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qptfm" podStartSLOduration=2.929389903 podStartE2EDuration="5.433927422s" podCreationTimestamp="2025-12-04 14:35:37 +0000 UTC" firstStartedPulling="2025-12-04 14:35:39.355977955 +0000 UTC m=+2296.424696170" lastFinishedPulling="2025-12-04 14:35:41.860515474 +0000 UTC m=+2298.929233689" observedRunningTime="2025-12-04 14:35:42.417437906 +0000 UTC m=+2299.486156141" watchObservedRunningTime="2025-12-04 14:35:42.433927422 +0000 UTC m=+2299.502645677" Dec 04 14:35:48 crc kubenswrapper[4715]: I1204 14:35:48.208270 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:48 crc kubenswrapper[4715]: I1204 14:35:48.208754 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:48 crc kubenswrapper[4715]: I1204 14:35:48.273454 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:48 crc kubenswrapper[4715]: I1204 14:35:48.516194 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:48 crc kubenswrapper[4715]: I1204 14:35:48.572740 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qptfm"] Dec 04 14:35:50 crc kubenswrapper[4715]: I1204 14:35:50.482344 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qptfm" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="registry-server" containerID="cri-o://a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095" gracePeriod=2 Dec 04 14:35:50 crc kubenswrapper[4715]: I1204 14:35:50.906069 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.032543 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-utilities\") pod \"c8244f07-d253-45fc-94af-09de9d42b8ce\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.032627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd4nl\" (UniqueName: \"kubernetes.io/projected/c8244f07-d253-45fc-94af-09de9d42b8ce-kube-api-access-jd4nl\") pod \"c8244f07-d253-45fc-94af-09de9d42b8ce\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.032852 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-catalog-content\") pod \"c8244f07-d253-45fc-94af-09de9d42b8ce\" (UID: \"c8244f07-d253-45fc-94af-09de9d42b8ce\") " Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.033425 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-utilities" (OuterVolumeSpecName: "utilities") pod "c8244f07-d253-45fc-94af-09de9d42b8ce" (UID: "c8244f07-d253-45fc-94af-09de9d42b8ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.039314 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8244f07-d253-45fc-94af-09de9d42b8ce-kube-api-access-jd4nl" (OuterVolumeSpecName: "kube-api-access-jd4nl") pod "c8244f07-d253-45fc-94af-09de9d42b8ce" (UID: "c8244f07-d253-45fc-94af-09de9d42b8ce"). InnerVolumeSpecName "kube-api-access-jd4nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.135740 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.135790 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd4nl\" (UniqueName: \"kubernetes.io/projected/c8244f07-d253-45fc-94af-09de9d42b8ce-kube-api-access-jd4nl\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.259025 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8244f07-d253-45fc-94af-09de9d42b8ce" (UID: "c8244f07-d253-45fc-94af-09de9d42b8ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.357974 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8244f07-d253-45fc-94af-09de9d42b8ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.509514 4715 generic.go:334] "Generic (PLEG): container finished" podID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerID="a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095" exitCode=0 Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.509600 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerDied","Data":"a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095"} Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.509642 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qptfm" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.509831 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qptfm" event={"ID":"c8244f07-d253-45fc-94af-09de9d42b8ce","Type":"ContainerDied","Data":"a35ccf830daa1fc95bcc6d1dd86cc84c3a8e19f4c1dee25564fdf9b742e3ba07"} Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.509862 4715 scope.go:117] "RemoveContainer" containerID="a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.530977 4715 scope.go:117] "RemoveContainer" containerID="082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.555246 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qptfm"] Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.562726 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qptfm"] Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.569696 4715 scope.go:117] "RemoveContainer" containerID="8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.595262 4715 scope.go:117] "RemoveContainer" containerID="a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095" Dec 04 14:35:51 crc kubenswrapper[4715]: E1204 14:35:51.595763 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095\": container with ID starting with a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095 not found: ID does not exist" containerID="a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.595813 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095"} err="failed to get container status \"a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095\": rpc error: code = NotFound desc = could not find container \"a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095\": container with ID starting with a8cc160fc744ab240293300f8a66dadee0b4ffb30bbd5ea97aa16313ae7c7095 not found: ID does not exist" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.595841 4715 scope.go:117] "RemoveContainer" containerID="082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1" Dec 04 14:35:51 crc kubenswrapper[4715]: E1204 14:35:51.596445 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1\": container with ID starting with 082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1 not found: ID does not exist" containerID="082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.596493 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1"} err="failed to get container status \"082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1\": rpc error: code = NotFound desc = could not find container \"082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1\": container with ID starting with 082fc5d50ef41f066bd53fef77ca704df51e97ab6bff6241bb7babf8bc6ae3f1 not found: ID does not exist" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.596521 4715 scope.go:117] "RemoveContainer" containerID="8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16" Dec 04 14:35:51 crc kubenswrapper[4715]: E1204 14:35:51.596796 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16\": container with ID starting with 8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16 not found: ID does not exist" containerID="8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16" Dec 04 14:35:51 crc kubenswrapper[4715]: I1204 14:35:51.596830 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16"} err="failed to get container status \"8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16\": rpc error: code = NotFound desc = could not find container \"8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16\": container with ID starting with 8beec4d1efc0074e12cf2bedfcbd1cec54219896368e862dc455e41b16f3fd16 not found: ID does not exist" Dec 04 14:35:53 crc kubenswrapper[4715]: I1204 14:35:53.191928 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" path="/var/lib/kubelet/pods/c8244f07-d253-45fc-94af-09de9d42b8ce/volumes" Dec 04 14:36:38 crc kubenswrapper[4715]: I1204 14:36:38.758801 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:36:38 crc kubenswrapper[4715]: I1204 14:36:38.759670 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:37:08 crc kubenswrapper[4715]: I1204 14:37:08.758406 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:37:08 crc kubenswrapper[4715]: I1204 14:37:08.758849 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:37:38 crc kubenswrapper[4715]: I1204 14:37:38.758556 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:37:38 crc kubenswrapper[4715]: I1204 14:37:38.758997 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:37:38 crc kubenswrapper[4715]: I1204 14:37:38.759065 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:37:38 crc kubenswrapper[4715]: I1204 14:37:38.759679 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:37:38 crc kubenswrapper[4715]: I1204 14:37:38.759736 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" gracePeriod=600 Dec 04 14:37:39 crc kubenswrapper[4715]: E1204 14:37:39.502839 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:37:39 crc kubenswrapper[4715]: I1204 14:37:39.591865 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" exitCode=0 Dec 04 14:37:39 crc kubenswrapper[4715]: I1204 14:37:39.591920 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a"} Dec 04 14:37:39 crc kubenswrapper[4715]: I1204 14:37:39.591958 4715 scope.go:117] "RemoveContainer" containerID="e52fdf637b0ee5a2201c3751e1bf6e2b0117b3148977de7bf1774674b74cb9be" Dec 04 14:37:39 crc kubenswrapper[4715]: I1204 14:37:39.592716 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:37:39 crc kubenswrapper[4715]: E1204 14:37:39.593069 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:37:50 crc kubenswrapper[4715]: I1204 14:37:50.181589 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:37:50 crc kubenswrapper[4715]: E1204 14:37:50.182299 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:38:05 crc kubenswrapper[4715]: I1204 14:38:05.180820 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:38:05 crc kubenswrapper[4715]: E1204 14:38:05.181662 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:38:20 crc kubenswrapper[4715]: I1204 14:38:20.181150 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:38:20 crc kubenswrapper[4715]: E1204 14:38:20.181982 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:38:35 crc kubenswrapper[4715]: I1204 14:38:35.181297 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:38:35 crc kubenswrapper[4715]: E1204 14:38:35.182167 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:38:48 crc kubenswrapper[4715]: I1204 14:38:48.181419 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:38:48 crc kubenswrapper[4715]: E1204 14:38:48.182720 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:39:03 crc kubenswrapper[4715]: I1204 14:39:03.188822 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:39:03 crc kubenswrapper[4715]: E1204 14:39:03.190245 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:39:14 crc kubenswrapper[4715]: I1204 14:39:14.181456 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:39:14 crc kubenswrapper[4715]: E1204 14:39:14.182343 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:39:29 crc kubenswrapper[4715]: I1204 14:39:29.180825 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:39:29 crc kubenswrapper[4715]: E1204 14:39:29.181709 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:39:33 crc kubenswrapper[4715]: E1204 14:39:33.802696 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a13c152_6be0_4261_ae2f_3b7b3458005a.slice/crio-conmon-624bcaaadd2f8fdaaeb88af1dc6b8e84c2ec3ea67fe06d924297e2cacf0737f7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a13c152_6be0_4261_ae2f_3b7b3458005a.slice/crio-624bcaaadd2f8fdaaeb88af1dc6b8e84c2ec3ea67fe06d924297e2cacf0737f7.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:39:34 crc kubenswrapper[4715]: I1204 14:39:34.728521 4715 generic.go:334] "Generic (PLEG): container finished" podID="4a13c152-6be0-4261-ae2f-3b7b3458005a" containerID="624bcaaadd2f8fdaaeb88af1dc6b8e84c2ec3ea67fe06d924297e2cacf0737f7" exitCode=0 Dec 04 14:39:34 crc kubenswrapper[4715]: I1204 14:39:34.728587 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" event={"ID":"4a13c152-6be0-4261-ae2f-3b7b3458005a","Type":"ContainerDied","Data":"624bcaaadd2f8fdaaeb88af1dc6b8e84c2ec3ea67fe06d924297e2cacf0737f7"} Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.174009 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.314828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjksl\" (UniqueName: \"kubernetes.io/projected/4a13c152-6be0-4261-ae2f-3b7b3458005a-kube-api-access-hjksl\") pod \"4a13c152-6be0-4261-ae2f-3b7b3458005a\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.314957 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-combined-ca-bundle\") pod \"4a13c152-6be0-4261-ae2f-3b7b3458005a\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.315036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-ssh-key\") pod \"4a13c152-6be0-4261-ae2f-3b7b3458005a\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.315110 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-secret-0\") pod \"4a13c152-6be0-4261-ae2f-3b7b3458005a\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.315172 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-inventory\") pod \"4a13c152-6be0-4261-ae2f-3b7b3458005a\" (UID: \"4a13c152-6be0-4261-ae2f-3b7b3458005a\") " Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.349280 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4a13c152-6be0-4261-ae2f-3b7b3458005a" (UID: "4a13c152-6be0-4261-ae2f-3b7b3458005a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.375432 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a13c152-6be0-4261-ae2f-3b7b3458005a-kube-api-access-hjksl" (OuterVolumeSpecName: "kube-api-access-hjksl") pod "4a13c152-6be0-4261-ae2f-3b7b3458005a" (UID: "4a13c152-6be0-4261-ae2f-3b7b3458005a"). InnerVolumeSpecName "kube-api-access-hjksl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.385352 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4a13c152-6be0-4261-ae2f-3b7b3458005a" (UID: "4a13c152-6be0-4261-ae2f-3b7b3458005a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.395566 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a13c152-6be0-4261-ae2f-3b7b3458005a" (UID: "4a13c152-6be0-4261-ae2f-3b7b3458005a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.408691 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-inventory" (OuterVolumeSpecName: "inventory") pod "4a13c152-6be0-4261-ae2f-3b7b3458005a" (UID: "4a13c152-6be0-4261-ae2f-3b7b3458005a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.417408 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.417485 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjksl\" (UniqueName: \"kubernetes.io/projected/4a13c152-6be0-4261-ae2f-3b7b3458005a-kube-api-access-hjksl\") on node \"crc\" DevicePath \"\"" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.417501 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.417511 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.417524 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4a13c152-6be0-4261-ae2f-3b7b3458005a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.747764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" event={"ID":"4a13c152-6be0-4261-ae2f-3b7b3458005a","Type":"ContainerDied","Data":"7f35719d53fac254e6ed9aba8e32ca096f577eb1cb79a19f49524a613ef0acdb"} Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.747810 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f35719d53fac254e6ed9aba8e32ca096f577eb1cb79a19f49524a613ef0acdb" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.747822 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.835902 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt"] Dec 04 14:39:36 crc kubenswrapper[4715]: E1204 14:39:36.836319 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="extract-content" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.836335 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="extract-content" Dec 04 14:39:36 crc kubenswrapper[4715]: E1204 14:39:36.836354 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a13c152-6be0-4261-ae2f-3b7b3458005a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.836361 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a13c152-6be0-4261-ae2f-3b7b3458005a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 14:39:36 crc kubenswrapper[4715]: E1204 14:39:36.836380 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="extract-utilities" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.836387 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="extract-utilities" Dec 04 14:39:36 crc kubenswrapper[4715]: E1204 14:39:36.836394 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="registry-server" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.836400 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="registry-server" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.836594 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a13c152-6be0-4261-ae2f-3b7b3458005a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.836626 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8244f07-d253-45fc-94af-09de9d42b8ce" containerName="registry-server" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.837736 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.840004 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.840261 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.840282 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.840490 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.841444 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.841471 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.841762 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.871604 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt"] Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925452 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925512 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925546 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925574 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925617 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925651 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925723 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkzmm\" (UniqueName: \"kubernetes.io/projected/df89956f-dd09-4cd1-bc58-d57f555d2870-kube-api-access-bkzmm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:36 crc kubenswrapper[4715]: I1204 14:39:36.925777 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.027702 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.027862 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.027966 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.028021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.028116 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.028214 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkzmm\" (UniqueName: \"kubernetes.io/projected/df89956f-dd09-4cd1-bc58-d57f555d2870-kube-api-access-bkzmm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.028301 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.028394 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.028497 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.029118 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.032476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.032669 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.033113 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.033675 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.034231 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.034311 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.035689 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.051081 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkzmm\" (UniqueName: \"kubernetes.io/projected/df89956f-dd09-4cd1-bc58-d57f555d2870-kube-api-access-bkzmm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gjztt\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.168814 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:39:37 crc kubenswrapper[4715]: I1204 14:39:37.779642 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt"] Dec 04 14:39:38 crc kubenswrapper[4715]: I1204 14:39:38.769687 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" event={"ID":"df89956f-dd09-4cd1-bc58-d57f555d2870","Type":"ContainerStarted","Data":"13d8984f058a510185d4adb4524969fa372296ece0b19ad939179de63da53520"} Dec 04 14:39:38 crc kubenswrapper[4715]: I1204 14:39:38.770052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" event={"ID":"df89956f-dd09-4cd1-bc58-d57f555d2870","Type":"ContainerStarted","Data":"d9f7c8d689bc15ba285337d68fd423868a572ed1fc7ed911fca59da45a289dc8"} Dec 04 14:39:38 crc kubenswrapper[4715]: I1204 14:39:38.791618 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" podStartSLOduration=2.37459318 podStartE2EDuration="2.791600988s" podCreationTimestamp="2025-12-04 14:39:36 +0000 UTC" firstStartedPulling="2025-12-04 14:39:37.795750779 +0000 UTC m=+2534.864468984" lastFinishedPulling="2025-12-04 14:39:38.212758577 +0000 UTC m=+2535.281476792" observedRunningTime="2025-12-04 14:39:38.78764432 +0000 UTC m=+2535.856362545" watchObservedRunningTime="2025-12-04 14:39:38.791600988 +0000 UTC m=+2535.860319193" Dec 04 14:39:42 crc kubenswrapper[4715]: I1204 14:39:42.181570 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:39:42 crc kubenswrapper[4715]: E1204 14:39:42.182431 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:39:54 crc kubenswrapper[4715]: I1204 14:39:54.180128 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:39:54 crc kubenswrapper[4715]: E1204 14:39:54.181150 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:40:08 crc kubenswrapper[4715]: I1204 14:40:08.181987 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:40:08 crc kubenswrapper[4715]: E1204 14:40:08.182867 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:40:20 crc kubenswrapper[4715]: I1204 14:40:20.389396 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:40:20 crc kubenswrapper[4715]: E1204 14:40:20.390238 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:40:35 crc kubenswrapper[4715]: I1204 14:40:35.180626 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:40:35 crc kubenswrapper[4715]: E1204 14:40:35.181595 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:40:39 crc kubenswrapper[4715]: I1204 14:40:39.364886 4715 scope.go:117] "RemoveContainer" containerID="803c1fbd18a8131c90011ded7c72d39019fd95782555e15fe4870512934e5196" Dec 04 14:40:39 crc kubenswrapper[4715]: I1204 14:40:39.394238 4715 scope.go:117] "RemoveContainer" containerID="a9886c892999ebefe8a112d2bf554a29a92d33078f01994f78a781eef4b2ae9e" Dec 04 14:40:39 crc kubenswrapper[4715]: I1204 14:40:39.444104 4715 scope.go:117] "RemoveContainer" containerID="ba303bcf87a5fa5904272869adc57b2963d551d8ef9d75911c6b3905ba988b72" Dec 04 14:40:48 crc kubenswrapper[4715]: I1204 14:40:48.180979 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:40:48 crc kubenswrapper[4715]: E1204 14:40:48.181855 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:41:00 crc kubenswrapper[4715]: I1204 14:41:00.181475 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:41:00 crc kubenswrapper[4715]: E1204 14:41:00.183271 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:41:12 crc kubenswrapper[4715]: I1204 14:41:12.180832 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:41:12 crc kubenswrapper[4715]: E1204 14:41:12.181964 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:41:27 crc kubenswrapper[4715]: I1204 14:41:27.180972 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:41:27 crc kubenswrapper[4715]: E1204 14:41:27.181809 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:41:39 crc kubenswrapper[4715]: I1204 14:41:39.181169 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:41:39 crc kubenswrapper[4715]: E1204 14:41:39.182115 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:41:50 crc kubenswrapper[4715]: I1204 14:41:50.181569 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:41:50 crc kubenswrapper[4715]: E1204 14:41:50.182420 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:42:04 crc kubenswrapper[4715]: I1204 14:42:04.181137 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:42:04 crc kubenswrapper[4715]: E1204 14:42:04.182274 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:42:15 crc kubenswrapper[4715]: I1204 14:42:15.182008 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:42:15 crc kubenswrapper[4715]: E1204 14:42:15.190075 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:42:16 crc kubenswrapper[4715]: I1204 14:42:16.470544 4715 generic.go:334] "Generic (PLEG): container finished" podID="df89956f-dd09-4cd1-bc58-d57f555d2870" containerID="13d8984f058a510185d4adb4524969fa372296ece0b19ad939179de63da53520" exitCode=0 Dec 04 14:42:16 crc kubenswrapper[4715]: I1204 14:42:16.470646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" event={"ID":"df89956f-dd09-4cd1-bc58-d57f555d2870","Type":"ContainerDied","Data":"13d8984f058a510185d4adb4524969fa372296ece0b19ad939179de63da53520"} Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.867414 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997077 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-0\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-1\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997210 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-inventory\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997288 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-combined-ca-bundle\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997320 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-extra-config-0\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997361 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-1\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997478 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkzmm\" (UniqueName: \"kubernetes.io/projected/df89956f-dd09-4cd1-bc58-d57f555d2870-kube-api-access-bkzmm\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997515 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-0\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:17 crc kubenswrapper[4715]: I1204 14:42:17.997566 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-ssh-key\") pod \"df89956f-dd09-4cd1-bc58-d57f555d2870\" (UID: \"df89956f-dd09-4cd1-bc58-d57f555d2870\") " Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.004388 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df89956f-dd09-4cd1-bc58-d57f555d2870-kube-api-access-bkzmm" (OuterVolumeSpecName: "kube-api-access-bkzmm") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "kube-api-access-bkzmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.004454 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.028107 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.028529 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-inventory" (OuterVolumeSpecName: "inventory") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.030107 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.034622 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.035242 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.036474 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.056561 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "df89956f-dd09-4cd1-bc58-d57f555d2870" (UID: "df89956f-dd09-4cd1-bc58-d57f555d2870"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102867 4715 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102912 4715 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102927 4715 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102941 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkzmm\" (UniqueName: \"kubernetes.io/projected/df89956f-dd09-4cd1-bc58-d57f555d2870-kube-api-access-bkzmm\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102955 4715 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102970 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.102987 4715 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.103008 4715 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.103027 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df89956f-dd09-4cd1-bc58-d57f555d2870-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.491516 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" event={"ID":"df89956f-dd09-4cd1-bc58-d57f555d2870","Type":"ContainerDied","Data":"d9f7c8d689bc15ba285337d68fd423868a572ed1fc7ed911fca59da45a289dc8"} Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.491926 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f7c8d689bc15ba285337d68fd423868a572ed1fc7ed911fca59da45a289dc8" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.491724 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gjztt" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.605144 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz"] Dec 04 14:42:18 crc kubenswrapper[4715]: E1204 14:42:18.605532 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df89956f-dd09-4cd1-bc58-d57f555d2870" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.605547 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="df89956f-dd09-4cd1-bc58-d57f555d2870" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.605781 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="df89956f-dd09-4cd1-bc58-d57f555d2870" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.606447 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.610271 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611486 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611561 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611584 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611609 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww8kg\" (UniqueName: \"kubernetes.io/projected/2fa80717-8afa-4d1a-ac4f-4b807445f638-kube-api-access-ww8kg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611630 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.611841 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.612021 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zzxcq" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.612202 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.612393 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.617879 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz"] Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.713517 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww8kg\" (UniqueName: \"kubernetes.io/projected/2fa80717-8afa-4d1a-ac4f-4b807445f638-kube-api-access-ww8kg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.713638 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.714592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.714690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.714733 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.714875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.714933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.718516 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.718760 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.719106 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.721632 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.728422 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.730348 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.730599 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww8kg\" (UniqueName: \"kubernetes.io/projected/2fa80717-8afa-4d1a-ac4f-4b807445f638-kube-api-access-ww8kg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:18 crc kubenswrapper[4715]: I1204 14:42:18.939434 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:42:19 crc kubenswrapper[4715]: I1204 14:42:19.459256 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz"] Dec 04 14:42:19 crc kubenswrapper[4715]: I1204 14:42:19.467581 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:42:19 crc kubenswrapper[4715]: I1204 14:42:19.501020 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" event={"ID":"2fa80717-8afa-4d1a-ac4f-4b807445f638","Type":"ContainerStarted","Data":"766b3acaee842d8e8fbeafdaf502bb3e6b24f08209c5dd3750193f75e32543b9"} Dec 04 14:42:21 crc kubenswrapper[4715]: I1204 14:42:21.522790 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" event={"ID":"2fa80717-8afa-4d1a-ac4f-4b807445f638","Type":"ContainerStarted","Data":"0c0e9ec63c577a1caf930b8015ed3482ceb30d749589489fc085116cae927d6e"} Dec 04 14:42:21 crc kubenswrapper[4715]: I1204 14:42:21.552771 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" podStartSLOduration=2.628140863 podStartE2EDuration="3.552748956s" podCreationTimestamp="2025-12-04 14:42:18 +0000 UTC" firstStartedPulling="2025-12-04 14:42:19.467245943 +0000 UTC m=+2696.535964158" lastFinishedPulling="2025-12-04 14:42:20.391854026 +0000 UTC m=+2697.460572251" observedRunningTime="2025-12-04 14:42:21.545534139 +0000 UTC m=+2698.614252354" watchObservedRunningTime="2025-12-04 14:42:21.552748956 +0000 UTC m=+2698.621467171" Dec 04 14:42:30 crc kubenswrapper[4715]: I1204 14:42:30.180688 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:42:30 crc kubenswrapper[4715]: E1204 14:42:30.181704 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:42:45 crc kubenswrapper[4715]: I1204 14:42:45.181248 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:42:45 crc kubenswrapper[4715]: I1204 14:42:45.774918 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"3bde09932246f62233cddfa1e31a9c959096be21967b6b67dfbfba81b6e806e4"} Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.592586 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f44sz"] Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.595924 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.614782 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f44sz"] Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.762410 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-utilities\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.762770 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zf2b\" (UniqueName: \"kubernetes.io/projected/656a380c-5ebd-4ff2-8321-49e3992adbcc-kube-api-access-7zf2b\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.762913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-catalog-content\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.864870 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-utilities\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.864953 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zf2b\" (UniqueName: \"kubernetes.io/projected/656a380c-5ebd-4ff2-8321-49e3992adbcc-kube-api-access-7zf2b\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.864986 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-catalog-content\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.865390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-utilities\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.865466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-catalog-content\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.894824 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zf2b\" (UniqueName: \"kubernetes.io/projected/656a380c-5ebd-4ff2-8321-49e3992adbcc-kube-api-access-7zf2b\") pod \"redhat-operators-f44sz\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:06 crc kubenswrapper[4715]: I1204 14:44:06.920152 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:07 crc kubenswrapper[4715]: I1204 14:44:07.415990 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f44sz"] Dec 04 14:44:08 crc kubenswrapper[4715]: I1204 14:44:08.177542 4715 generic.go:334] "Generic (PLEG): container finished" podID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerID="25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b" exitCode=0 Dec 04 14:44:08 crc kubenswrapper[4715]: I1204 14:44:08.177685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerDied","Data":"25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b"} Dec 04 14:44:08 crc kubenswrapper[4715]: I1204 14:44:08.177897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerStarted","Data":"dbd56ff6d43048077f12433aad282e2bd2d190b47ba3d4660fab60b2a248614c"} Dec 04 14:44:09 crc kubenswrapper[4715]: I1204 14:44:09.200952 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerStarted","Data":"25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369"} Dec 04 14:44:10 crc kubenswrapper[4715]: I1204 14:44:10.211958 4715 generic.go:334] "Generic (PLEG): container finished" podID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerID="25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369" exitCode=0 Dec 04 14:44:10 crc kubenswrapper[4715]: I1204 14:44:10.212001 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerDied","Data":"25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369"} Dec 04 14:44:12 crc kubenswrapper[4715]: I1204 14:44:12.231388 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerStarted","Data":"9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa"} Dec 04 14:44:14 crc kubenswrapper[4715]: I1204 14:44:14.268302 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f44sz" podStartSLOduration=5.651503917 podStartE2EDuration="8.268280812s" podCreationTimestamp="2025-12-04 14:44:06 +0000 UTC" firstStartedPulling="2025-12-04 14:44:08.18107774 +0000 UTC m=+2805.249795955" lastFinishedPulling="2025-12-04 14:44:10.797854635 +0000 UTC m=+2807.866572850" observedRunningTime="2025-12-04 14:44:14.264975992 +0000 UTC m=+2811.333694207" watchObservedRunningTime="2025-12-04 14:44:14.268280812 +0000 UTC m=+2811.336999027" Dec 04 14:44:16 crc kubenswrapper[4715]: I1204 14:44:16.920519 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:16 crc kubenswrapper[4715]: I1204 14:44:16.920940 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:16 crc kubenswrapper[4715]: I1204 14:44:16.970231 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:17 crc kubenswrapper[4715]: I1204 14:44:17.314164 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:17 crc kubenswrapper[4715]: I1204 14:44:17.371520 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f44sz"] Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.287361 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f44sz" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="registry-server" containerID="cri-o://9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa" gracePeriod=2 Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.708783 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.829570 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-utilities\") pod \"656a380c-5ebd-4ff2-8321-49e3992adbcc\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.829897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zf2b\" (UniqueName: \"kubernetes.io/projected/656a380c-5ebd-4ff2-8321-49e3992adbcc-kube-api-access-7zf2b\") pod \"656a380c-5ebd-4ff2-8321-49e3992adbcc\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.830045 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-catalog-content\") pod \"656a380c-5ebd-4ff2-8321-49e3992adbcc\" (UID: \"656a380c-5ebd-4ff2-8321-49e3992adbcc\") " Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.833251 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-utilities" (OuterVolumeSpecName: "utilities") pod "656a380c-5ebd-4ff2-8321-49e3992adbcc" (UID: "656a380c-5ebd-4ff2-8321-49e3992adbcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.836604 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656a380c-5ebd-4ff2-8321-49e3992adbcc-kube-api-access-7zf2b" (OuterVolumeSpecName: "kube-api-access-7zf2b") pod "656a380c-5ebd-4ff2-8321-49e3992adbcc" (UID: "656a380c-5ebd-4ff2-8321-49e3992adbcc"). InnerVolumeSpecName "kube-api-access-7zf2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.932431 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.932465 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zf2b\" (UniqueName: \"kubernetes.io/projected/656a380c-5ebd-4ff2-8321-49e3992adbcc-kube-api-access-7zf2b\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:19 crc kubenswrapper[4715]: I1204 14:44:19.949082 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "656a380c-5ebd-4ff2-8321-49e3992adbcc" (UID: "656a380c-5ebd-4ff2-8321-49e3992adbcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.034390 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656a380c-5ebd-4ff2-8321-49e3992adbcc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.298416 4715 generic.go:334] "Generic (PLEG): container finished" podID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerID="9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa" exitCode=0 Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.298476 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerDied","Data":"9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa"} Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.298504 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f44sz" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.298538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f44sz" event={"ID":"656a380c-5ebd-4ff2-8321-49e3992adbcc","Type":"ContainerDied","Data":"dbd56ff6d43048077f12433aad282e2bd2d190b47ba3d4660fab60b2a248614c"} Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.298559 4715 scope.go:117] "RemoveContainer" containerID="9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.331368 4715 scope.go:117] "RemoveContainer" containerID="25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.345479 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f44sz"] Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.354706 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f44sz"] Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.381374 4715 scope.go:117] "RemoveContainer" containerID="25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.401215 4715 scope.go:117] "RemoveContainer" containerID="9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa" Dec 04 14:44:20 crc kubenswrapper[4715]: E1204 14:44:20.401691 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa\": container with ID starting with 9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa not found: ID does not exist" containerID="9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.401750 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa"} err="failed to get container status \"9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa\": rpc error: code = NotFound desc = could not find container \"9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa\": container with ID starting with 9f814a8c1666f0a3e36a8369cb01c8a741e2f9d03fb6ffe7136974eb1186a8fa not found: ID does not exist" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.401792 4715 scope.go:117] "RemoveContainer" containerID="25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369" Dec 04 14:44:20 crc kubenswrapper[4715]: E1204 14:44:20.402283 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369\": container with ID starting with 25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369 not found: ID does not exist" containerID="25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.402314 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369"} err="failed to get container status \"25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369\": rpc error: code = NotFound desc = could not find container \"25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369\": container with ID starting with 25a94cb9feb1586803120a71bc5e4eba7ca45f51ebe2883a079d22cd65f84369 not found: ID does not exist" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.402337 4715 scope.go:117] "RemoveContainer" containerID="25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b" Dec 04 14:44:20 crc kubenswrapper[4715]: E1204 14:44:20.402550 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b\": container with ID starting with 25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b not found: ID does not exist" containerID="25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b" Dec 04 14:44:20 crc kubenswrapper[4715]: I1204 14:44:20.402575 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b"} err="failed to get container status \"25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b\": rpc error: code = NotFound desc = could not find container \"25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b\": container with ID starting with 25cb0fd20bbaba25836b85d8778623a39dcb6b5feba608cfd08c2b26fed0165b not found: ID does not exist" Dec 04 14:44:21 crc kubenswrapper[4715]: I1204 14:44:21.190549 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" path="/var/lib/kubelet/pods/656a380c-5ebd-4ff2-8321-49e3992adbcc/volumes" Dec 04 14:44:35 crc kubenswrapper[4715]: I1204 14:44:35.442473 4715 generic.go:334] "Generic (PLEG): container finished" podID="2fa80717-8afa-4d1a-ac4f-4b807445f638" containerID="0c0e9ec63c577a1caf930b8015ed3482ceb30d749589489fc085116cae927d6e" exitCode=0 Dec 04 14:44:35 crc kubenswrapper[4715]: I1204 14:44:35.443059 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" event={"ID":"2fa80717-8afa-4d1a-ac4f-4b807445f638","Type":"ContainerDied","Data":"0c0e9ec63c577a1caf930b8015ed3482ceb30d749589489fc085116cae927d6e"} Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.871344 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967407 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-telemetry-combined-ca-bundle\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967505 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww8kg\" (UniqueName: \"kubernetes.io/projected/2fa80717-8afa-4d1a-ac4f-4b807445f638-kube-api-access-ww8kg\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967565 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-2\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967600 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-0\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967694 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ssh-key\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-inventory\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.967811 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-1\") pod \"2fa80717-8afa-4d1a-ac4f-4b807445f638\" (UID: \"2fa80717-8afa-4d1a-ac4f-4b807445f638\") " Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.978343 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa80717-8afa-4d1a-ac4f-4b807445f638-kube-api-access-ww8kg" (OuterVolumeSpecName: "kube-api-access-ww8kg") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "kube-api-access-ww8kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.978863 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.995286 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.996991 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:44:36 crc kubenswrapper[4715]: I1204 14:44:36.999229 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.000962 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-inventory" (OuterVolumeSpecName: "inventory") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.004360 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2fa80717-8afa-4d1a-ac4f-4b807445f638" (UID: "2fa80717-8afa-4d1a-ac4f-4b807445f638"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070201 4715 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070240 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww8kg\" (UniqueName: \"kubernetes.io/projected/2fa80717-8afa-4d1a-ac4f-4b807445f638-kube-api-access-ww8kg\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070249 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070259 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070271 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070281 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.070289 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2fa80717-8afa-4d1a-ac4f-4b807445f638-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.460410 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" event={"ID":"2fa80717-8afa-4d1a-ac4f-4b807445f638","Type":"ContainerDied","Data":"766b3acaee842d8e8fbeafdaf502bb3e6b24f08209c5dd3750193f75e32543b9"} Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.460452 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="766b3acaee842d8e8fbeafdaf502bb3e6b24f08209c5dd3750193f75e32543b9" Dec 04 14:44:37 crc kubenswrapper[4715]: I1204 14:44:37.460503 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.145783 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8"] Dec 04 14:45:00 crc kubenswrapper[4715]: E1204 14:45:00.146709 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="registry-server" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.146723 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="registry-server" Dec 04 14:45:00 crc kubenswrapper[4715]: E1204 14:45:00.146742 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa80717-8afa-4d1a-ac4f-4b807445f638" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.146749 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa80717-8afa-4d1a-ac4f-4b807445f638" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 14:45:00 crc kubenswrapper[4715]: E1204 14:45:00.146762 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="extract-content" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.146768 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="extract-content" Dec 04 14:45:00 crc kubenswrapper[4715]: E1204 14:45:00.146787 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="extract-utilities" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.146794 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="extract-utilities" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.147030 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="656a380c-5ebd-4ff2-8321-49e3992adbcc" containerName="registry-server" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.147069 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa80717-8afa-4d1a-ac4f-4b807445f638" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.147852 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.149990 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.150218 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.168677 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8"] Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.233774 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-secret-volume\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.234456 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-config-volume\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.234534 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptdsh\" (UniqueName: \"kubernetes.io/projected/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-kube-api-access-ptdsh\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.336423 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-secret-volume\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.336895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-config-volume\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.337042 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptdsh\" (UniqueName: \"kubernetes.io/projected/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-kube-api-access-ptdsh\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.337840 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-config-volume\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.344517 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-secret-volume\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.361659 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptdsh\" (UniqueName: \"kubernetes.io/projected/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-kube-api-access-ptdsh\") pod \"collect-profiles-29414325-hbjg8\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.469307 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:00 crc kubenswrapper[4715]: I1204 14:45:00.957942 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8"] Dec 04 14:45:01 crc kubenswrapper[4715]: I1204 14:45:01.682708 4715 generic.go:334] "Generic (PLEG): container finished" podID="1ea91ef7-37a2-4f7e-8ab6-c613473b1573" containerID="49d11132687af4a9685d59720ed6fd4bcde34b648d2653fa8a639e200b37286a" exitCode=0 Dec 04 14:45:01 crc kubenswrapper[4715]: I1204 14:45:01.682972 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" event={"ID":"1ea91ef7-37a2-4f7e-8ab6-c613473b1573","Type":"ContainerDied","Data":"49d11132687af4a9685d59720ed6fd4bcde34b648d2653fa8a639e200b37286a"} Dec 04 14:45:01 crc kubenswrapper[4715]: I1204 14:45:01.683010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" event={"ID":"1ea91ef7-37a2-4f7e-8ab6-c613473b1573","Type":"ContainerStarted","Data":"8a529de8b743e9b00fb0670c0cd7fb969b3025ad27d0a17bd45a2a832891c90d"} Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.076643 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.190757 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-config-volume\") pod \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.191694 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-secret-volume\") pod \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.192139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptdsh\" (UniqueName: \"kubernetes.io/projected/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-kube-api-access-ptdsh\") pod \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\" (UID: \"1ea91ef7-37a2-4f7e-8ab6-c613473b1573\") " Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.193866 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-config-volume" (OuterVolumeSpecName: "config-volume") pod "1ea91ef7-37a2-4f7e-8ab6-c613473b1573" (UID: "1ea91ef7-37a2-4f7e-8ab6-c613473b1573"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.200955 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1ea91ef7-37a2-4f7e-8ab6-c613473b1573" (UID: "1ea91ef7-37a2-4f7e-8ab6-c613473b1573"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.201822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-kube-api-access-ptdsh" (OuterVolumeSpecName: "kube-api-access-ptdsh") pod "1ea91ef7-37a2-4f7e-8ab6-c613473b1573" (UID: "1ea91ef7-37a2-4f7e-8ab6-c613473b1573"). InnerVolumeSpecName "kube-api-access-ptdsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.299543 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptdsh\" (UniqueName: \"kubernetes.io/projected/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-kube-api-access-ptdsh\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.299588 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.299601 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ea91ef7-37a2-4f7e-8ab6-c613473b1573-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.705440 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" event={"ID":"1ea91ef7-37a2-4f7e-8ab6-c613473b1573","Type":"ContainerDied","Data":"8a529de8b743e9b00fb0670c0cd7fb969b3025ad27d0a17bd45a2a832891c90d"} Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.705752 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a529de8b743e9b00fb0670c0cd7fb969b3025ad27d0a17bd45a2a832891c90d" Dec 04 14:45:03 crc kubenswrapper[4715]: I1204 14:45:03.706552 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-hbjg8" Dec 04 14:45:04 crc kubenswrapper[4715]: I1204 14:45:04.159079 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn"] Dec 04 14:45:04 crc kubenswrapper[4715]: I1204 14:45:04.169142 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-b9rwn"] Dec 04 14:45:05 crc kubenswrapper[4715]: I1204 14:45:05.202848 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d8f924-0487-4f2f-8241-4dfe2c51fb01" path="/var/lib/kubelet/pods/c8d8f924-0487-4f2f-8241-4dfe2c51fb01/volumes" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.312129 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p9wld"] Dec 04 14:45:07 crc kubenswrapper[4715]: E1204 14:45:07.312960 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea91ef7-37a2-4f7e-8ab6-c613473b1573" containerName="collect-profiles" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.312976 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea91ef7-37a2-4f7e-8ab6-c613473b1573" containerName="collect-profiles" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.313245 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea91ef7-37a2-4f7e-8ab6-c613473b1573" containerName="collect-profiles" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.316298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.357857 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9wld"] Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.408523 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-catalog-content\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.408595 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-utilities\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.408685 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfspp\" (UniqueName: \"kubernetes.io/projected/9eec2b14-aeab-40fa-b46c-738009ef93c9-kube-api-access-rfspp\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.511640 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-catalog-content\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.511812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-utilities\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.511963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfspp\" (UniqueName: \"kubernetes.io/projected/9eec2b14-aeab-40fa-b46c-738009ef93c9-kube-api-access-rfspp\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.513120 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-catalog-content\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.513332 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-utilities\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.535214 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfspp\" (UniqueName: \"kubernetes.io/projected/9eec2b14-aeab-40fa-b46c-738009ef93c9-kube-api-access-rfspp\") pod \"redhat-marketplace-p9wld\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:07 crc kubenswrapper[4715]: I1204 14:45:07.640765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:08 crc kubenswrapper[4715]: I1204 14:45:08.091243 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9wld"] Dec 04 14:45:08 crc kubenswrapper[4715]: I1204 14:45:08.755813 4715 generic.go:334] "Generic (PLEG): container finished" podID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerID="a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac" exitCode=0 Dec 04 14:45:08 crc kubenswrapper[4715]: I1204 14:45:08.755895 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9wld" event={"ID":"9eec2b14-aeab-40fa-b46c-738009ef93c9","Type":"ContainerDied","Data":"a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac"} Dec 04 14:45:08 crc kubenswrapper[4715]: I1204 14:45:08.756106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9wld" event={"ID":"9eec2b14-aeab-40fa-b46c-738009ef93c9","Type":"ContainerStarted","Data":"eca0b1ef0f0c396dc85918a018e57aaf31ae5bcf09306ceaf5886855652e6f8b"} Dec 04 14:45:08 crc kubenswrapper[4715]: I1204 14:45:08.758580 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:45:08 crc kubenswrapper[4715]: I1204 14:45:08.758630 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:45:10 crc kubenswrapper[4715]: I1204 14:45:10.779893 4715 generic.go:334] "Generic (PLEG): container finished" podID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerID="29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440" exitCode=0 Dec 04 14:45:10 crc kubenswrapper[4715]: I1204 14:45:10.780139 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9wld" event={"ID":"9eec2b14-aeab-40fa-b46c-738009ef93c9","Type":"ContainerDied","Data":"29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440"} Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.689639 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t45vn"] Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.692673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.699118 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t45vn"] Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.800837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9wld" event={"ID":"9eec2b14-aeab-40fa-b46c-738009ef93c9","Type":"ContainerStarted","Data":"bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1"} Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.814191 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-catalog-content\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.814534 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-utilities\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.814659 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcr54\" (UniqueName: \"kubernetes.io/projected/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-kube-api-access-lcr54\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.825903 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p9wld" podStartSLOduration=2.270359889 podStartE2EDuration="5.825883782s" podCreationTimestamp="2025-12-04 14:45:07 +0000 UTC" firstStartedPulling="2025-12-04 14:45:08.758605901 +0000 UTC m=+2865.827324136" lastFinishedPulling="2025-12-04 14:45:12.314129814 +0000 UTC m=+2869.382848029" observedRunningTime="2025-12-04 14:45:12.821646366 +0000 UTC m=+2869.890364581" watchObservedRunningTime="2025-12-04 14:45:12.825883782 +0000 UTC m=+2869.894601997" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.916412 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-catalog-content\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.916529 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-utilities\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.916573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcr54\" (UniqueName: \"kubernetes.io/projected/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-kube-api-access-lcr54\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.916999 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-catalog-content\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.917111 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-utilities\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:12 crc kubenswrapper[4715]: I1204 14:45:12.954074 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcr54\" (UniqueName: \"kubernetes.io/projected/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-kube-api-access-lcr54\") pod \"community-operators-t45vn\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:13 crc kubenswrapper[4715]: I1204 14:45:13.012507 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:13 crc kubenswrapper[4715]: I1204 14:45:13.631269 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t45vn"] Dec 04 14:45:13 crc kubenswrapper[4715]: I1204 14:45:13.814300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t45vn" event={"ID":"cd1045fb-9f78-48a7-b342-4fddfad4f8ca","Type":"ContainerStarted","Data":"e486d2e645038b93323c4a291844776579494e94081e7073b20187d240685748"} Dec 04 14:45:14 crc kubenswrapper[4715]: I1204 14:45:14.823483 4715 generic.go:334] "Generic (PLEG): container finished" podID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerID="19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd" exitCode=0 Dec 04 14:45:14 crc kubenswrapper[4715]: I1204 14:45:14.823548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t45vn" event={"ID":"cd1045fb-9f78-48a7-b342-4fddfad4f8ca","Type":"ContainerDied","Data":"19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd"} Dec 04 14:45:17 crc kubenswrapper[4715]: I1204 14:45:17.642003 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:17 crc kubenswrapper[4715]: I1204 14:45:17.642976 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:17 crc kubenswrapper[4715]: I1204 14:45:17.744528 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:17 crc kubenswrapper[4715]: I1204 14:45:17.849696 4715 generic.go:334] "Generic (PLEG): container finished" podID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerID="69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083" exitCode=0 Dec 04 14:45:17 crc kubenswrapper[4715]: I1204 14:45:17.849799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t45vn" event={"ID":"cd1045fb-9f78-48a7-b342-4fddfad4f8ca","Type":"ContainerDied","Data":"69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083"} Dec 04 14:45:17 crc kubenswrapper[4715]: I1204 14:45:17.921584 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:18 crc kubenswrapper[4715]: I1204 14:45:18.862731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t45vn" event={"ID":"cd1045fb-9f78-48a7-b342-4fddfad4f8ca","Type":"ContainerStarted","Data":"8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1"} Dec 04 14:45:19 crc kubenswrapper[4715]: I1204 14:45:19.888352 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9wld"] Dec 04 14:45:19 crc kubenswrapper[4715]: I1204 14:45:19.892419 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t45vn" podStartSLOduration=4.126506987 podStartE2EDuration="7.89239906s" podCreationTimestamp="2025-12-04 14:45:12 +0000 UTC" firstStartedPulling="2025-12-04 14:45:14.825977351 +0000 UTC m=+2871.894695566" lastFinishedPulling="2025-12-04 14:45:18.591869424 +0000 UTC m=+2875.660587639" observedRunningTime="2025-12-04 14:45:19.88873747 +0000 UTC m=+2876.957455685" watchObservedRunningTime="2025-12-04 14:45:19.89239906 +0000 UTC m=+2876.961117275" Dec 04 14:45:20 crc kubenswrapper[4715]: I1204 14:45:20.956595 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p9wld" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="registry-server" containerID="cri-o://bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1" gracePeriod=2 Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.430532 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.452163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-catalog-content\") pod \"9eec2b14-aeab-40fa-b46c-738009ef93c9\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.452219 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-utilities\") pod \"9eec2b14-aeab-40fa-b46c-738009ef93c9\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.452373 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfspp\" (UniqueName: \"kubernetes.io/projected/9eec2b14-aeab-40fa-b46c-738009ef93c9-kube-api-access-rfspp\") pod \"9eec2b14-aeab-40fa-b46c-738009ef93c9\" (UID: \"9eec2b14-aeab-40fa-b46c-738009ef93c9\") " Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.455742 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-utilities" (OuterVolumeSpecName: "utilities") pod "9eec2b14-aeab-40fa-b46c-738009ef93c9" (UID: "9eec2b14-aeab-40fa-b46c-738009ef93c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.458455 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eec2b14-aeab-40fa-b46c-738009ef93c9-kube-api-access-rfspp" (OuterVolumeSpecName: "kube-api-access-rfspp") pod "9eec2b14-aeab-40fa-b46c-738009ef93c9" (UID: "9eec2b14-aeab-40fa-b46c-738009ef93c9"). InnerVolumeSpecName "kube-api-access-rfspp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.475490 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9eec2b14-aeab-40fa-b46c-738009ef93c9" (UID: "9eec2b14-aeab-40fa-b46c-738009ef93c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.554102 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfspp\" (UniqueName: \"kubernetes.io/projected/9eec2b14-aeab-40fa-b46c-738009ef93c9-kube-api-access-rfspp\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.554136 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.554145 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eec2b14-aeab-40fa-b46c-738009ef93c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.966772 4715 generic.go:334] "Generic (PLEG): container finished" podID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerID="bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1" exitCode=0 Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.966817 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9wld" event={"ID":"9eec2b14-aeab-40fa-b46c-738009ef93c9","Type":"ContainerDied","Data":"bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1"} Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.966845 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9wld" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.966882 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9wld" event={"ID":"9eec2b14-aeab-40fa-b46c-738009ef93c9","Type":"ContainerDied","Data":"eca0b1ef0f0c396dc85918a018e57aaf31ae5bcf09306ceaf5886855652e6f8b"} Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.966907 4715 scope.go:117] "RemoveContainer" containerID="bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1" Dec 04 14:45:21 crc kubenswrapper[4715]: I1204 14:45:21.988998 4715 scope.go:117] "RemoveContainer" containerID="29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.006697 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9wld"] Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.014872 4715 scope.go:117] "RemoveContainer" containerID="a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.016094 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9wld"] Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.059733 4715 scope.go:117] "RemoveContainer" containerID="bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1" Dec 04 14:45:22 crc kubenswrapper[4715]: E1204 14:45:22.060266 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1\": container with ID starting with bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1 not found: ID does not exist" containerID="bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.060299 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1"} err="failed to get container status \"bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1\": rpc error: code = NotFound desc = could not find container \"bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1\": container with ID starting with bbf305e916a68c8da591bbf47bb4dce081c1c068d9cb35106279fb7a80c572d1 not found: ID does not exist" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.060319 4715 scope.go:117] "RemoveContainer" containerID="29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440" Dec 04 14:45:22 crc kubenswrapper[4715]: E1204 14:45:22.060590 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440\": container with ID starting with 29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440 not found: ID does not exist" containerID="29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.060605 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440"} err="failed to get container status \"29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440\": rpc error: code = NotFound desc = could not find container \"29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440\": container with ID starting with 29075533d5f348751ecfee2d652a7c0fbbb7f4fc113cdb015ecee2246a9c4440 not found: ID does not exist" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.060616 4715 scope.go:117] "RemoveContainer" containerID="a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac" Dec 04 14:45:22 crc kubenswrapper[4715]: E1204 14:45:22.060994 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac\": container with ID starting with a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac not found: ID does not exist" containerID="a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac" Dec 04 14:45:22 crc kubenswrapper[4715]: I1204 14:45:22.061023 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac"} err="failed to get container status \"a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac\": rpc error: code = NotFound desc = could not find container \"a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac\": container with ID starting with a064f71ce8b7da552ae2bebf597a486d57852dd6322657341a4d9004f01940ac not found: ID does not exist" Dec 04 14:45:23 crc kubenswrapper[4715]: I1204 14:45:23.013451 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:23 crc kubenswrapper[4715]: I1204 14:45:23.013809 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:23 crc kubenswrapper[4715]: I1204 14:45:23.060762 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:23 crc kubenswrapper[4715]: I1204 14:45:23.192612 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" path="/var/lib/kubelet/pods/9eec2b14-aeab-40fa-b46c-738009ef93c9/volumes" Dec 04 14:45:24 crc kubenswrapper[4715]: I1204 14:45:24.039506 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:25 crc kubenswrapper[4715]: I1204 14:45:25.168466 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t45vn"] Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.014238 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t45vn" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="registry-server" containerID="cri-o://8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1" gracePeriod=2 Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.460937 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.517144 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-catalog-content\") pod \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.517269 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcr54\" (UniqueName: \"kubernetes.io/projected/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-kube-api-access-lcr54\") pod \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.517302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-utilities\") pod \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\" (UID: \"cd1045fb-9f78-48a7-b342-4fddfad4f8ca\") " Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.524009 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-kube-api-access-lcr54" (OuterVolumeSpecName: "kube-api-access-lcr54") pod "cd1045fb-9f78-48a7-b342-4fddfad4f8ca" (UID: "cd1045fb-9f78-48a7-b342-4fddfad4f8ca"). InnerVolumeSpecName "kube-api-access-lcr54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.525237 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-utilities" (OuterVolumeSpecName: "utilities") pod "cd1045fb-9f78-48a7-b342-4fddfad4f8ca" (UID: "cd1045fb-9f78-48a7-b342-4fddfad4f8ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.574480 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd1045fb-9f78-48a7-b342-4fddfad4f8ca" (UID: "cd1045fb-9f78-48a7-b342-4fddfad4f8ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.619696 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcr54\" (UniqueName: \"kubernetes.io/projected/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-kube-api-access-lcr54\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.619745 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:26 crc kubenswrapper[4715]: I1204 14:45:26.619755 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd1045fb-9f78-48a7-b342-4fddfad4f8ca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.026214 4715 generic.go:334] "Generic (PLEG): container finished" podID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerID="8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1" exitCode=0 Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.026304 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t45vn" event={"ID":"cd1045fb-9f78-48a7-b342-4fddfad4f8ca","Type":"ContainerDied","Data":"8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1"} Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.026340 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t45vn" event={"ID":"cd1045fb-9f78-48a7-b342-4fddfad4f8ca","Type":"ContainerDied","Data":"e486d2e645038b93323c4a291844776579494e94081e7073b20187d240685748"} Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.026362 4715 scope.go:117] "RemoveContainer" containerID="8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.026618 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t45vn" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.048203 4715 scope.go:117] "RemoveContainer" containerID="69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.061758 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t45vn"] Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.073268 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t45vn"] Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.082282 4715 scope.go:117] "RemoveContainer" containerID="19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.121028 4715 scope.go:117] "RemoveContainer" containerID="8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1" Dec 04 14:45:27 crc kubenswrapper[4715]: E1204 14:45:27.121509 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1\": container with ID starting with 8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1 not found: ID does not exist" containerID="8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.121554 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1"} err="failed to get container status \"8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1\": rpc error: code = NotFound desc = could not find container \"8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1\": container with ID starting with 8112da537f044b671fce94063865a54268adc2d73e0ff621a8f2f9199f48deb1 not found: ID does not exist" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.121584 4715 scope.go:117] "RemoveContainer" containerID="69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083" Dec 04 14:45:27 crc kubenswrapper[4715]: E1204 14:45:27.121846 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083\": container with ID starting with 69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083 not found: ID does not exist" containerID="69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.121888 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083"} err="failed to get container status \"69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083\": rpc error: code = NotFound desc = could not find container \"69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083\": container with ID starting with 69def442c4754b83482e550803a9e48322de524677bd5e916e6934befb4e9083 not found: ID does not exist" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.121904 4715 scope.go:117] "RemoveContainer" containerID="19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd" Dec 04 14:45:27 crc kubenswrapper[4715]: E1204 14:45:27.122448 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd\": container with ID starting with 19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd not found: ID does not exist" containerID="19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.122469 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd"} err="failed to get container status \"19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd\": rpc error: code = NotFound desc = could not find container \"19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd\": container with ID starting with 19c0a74e1a9bae2c86ac6a06dc3cba2de90d77c173a398ead6333e2b170f73fd not found: ID does not exist" Dec 04 14:45:27 crc kubenswrapper[4715]: I1204 14:45:27.195140 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" path="/var/lib/kubelet/pods/cd1045fb-9f78-48a7-b342-4fddfad4f8ca/volumes" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.462330 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 14:45:36 crc kubenswrapper[4715]: E1204 14:45:36.463404 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="registry-server" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463423 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="registry-server" Dec 04 14:45:36 crc kubenswrapper[4715]: E1204 14:45:36.463444 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="extract-utilities" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463453 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="extract-utilities" Dec 04 14:45:36 crc kubenswrapper[4715]: E1204 14:45:36.463473 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="extract-content" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463479 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="extract-content" Dec 04 14:45:36 crc kubenswrapper[4715]: E1204 14:45:36.463505 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="extract-utilities" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463511 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="extract-utilities" Dec 04 14:45:36 crc kubenswrapper[4715]: E1204 14:45:36.463531 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="registry-server" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463538 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="registry-server" Dec 04 14:45:36 crc kubenswrapper[4715]: E1204 14:45:36.463558 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="extract-content" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463565 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="extract-content" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463781 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1045fb-9f78-48a7-b342-4fddfad4f8ca" containerName="registry-server" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.463798 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eec2b14-aeab-40fa-b46c-738009ef93c9" containerName="registry-server" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.464713 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.466568 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.467225 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.467363 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-h87fs" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.468481 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.481698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.532109 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.532372 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-config-data\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.532421 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634219 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-452mv\" (UniqueName: \"kubernetes.io/projected/1314b612-6502-4cef-ade7-f05971b781a2-kube-api-access-452mv\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634293 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634472 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634600 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-config-data\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.634961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.635007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.635097 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.635821 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-config-data\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.636296 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.642158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737387 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737537 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737721 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-452mv\" (UniqueName: \"kubernetes.io/projected/1314b612-6502-4cef-ade7-f05971b781a2-kube-api-access-452mv\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.737992 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.738201 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.742571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.744881 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.757914 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-452mv\" (UniqueName: \"kubernetes.io/projected/1314b612-6502-4cef-ade7-f05971b781a2-kube-api-access-452mv\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.776779 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " pod="openstack/tempest-tests-tempest" Dec 04 14:45:36 crc kubenswrapper[4715]: I1204 14:45:36.802014 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 14:45:37 crc kubenswrapper[4715]: I1204 14:45:37.287217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 14:45:37 crc kubenswrapper[4715]: W1204 14:45:37.292319 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1314b612_6502_4cef_ade7_f05971b781a2.slice/crio-dc5e43ecc47a9f01960a8abb1e3c5a853eea68cd44aac2d3ae8577d8bb3415a7 WatchSource:0}: Error finding container dc5e43ecc47a9f01960a8abb1e3c5a853eea68cd44aac2d3ae8577d8bb3415a7: Status 404 returned error can't find the container with id dc5e43ecc47a9f01960a8abb1e3c5a853eea68cd44aac2d3ae8577d8bb3415a7 Dec 04 14:45:38 crc kubenswrapper[4715]: I1204 14:45:38.131160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1314b612-6502-4cef-ade7-f05971b781a2","Type":"ContainerStarted","Data":"dc5e43ecc47a9f01960a8abb1e3c5a853eea68cd44aac2d3ae8577d8bb3415a7"} Dec 04 14:45:38 crc kubenswrapper[4715]: I1204 14:45:38.759051 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:45:38 crc kubenswrapper[4715]: I1204 14:45:38.759439 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:45:39 crc kubenswrapper[4715]: I1204 14:45:39.606680 4715 scope.go:117] "RemoveContainer" containerID="ebf34ca00017234d68440996e44dcac5207890b1907b63eb25d0d6e56fd51840" Dec 04 14:46:08 crc kubenswrapper[4715]: I1204 14:46:08.758364 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:46:08 crc kubenswrapper[4715]: I1204 14:46:08.758909 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:46:08 crc kubenswrapper[4715]: I1204 14:46:08.758954 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:46:08 crc kubenswrapper[4715]: I1204 14:46:08.759764 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3bde09932246f62233cddfa1e31a9c959096be21967b6b67dfbfba81b6e806e4"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:46:08 crc kubenswrapper[4715]: I1204 14:46:08.759829 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://3bde09932246f62233cddfa1e31a9c959096be21967b6b67dfbfba81b6e806e4" gracePeriod=600 Dec 04 14:46:10 crc kubenswrapper[4715]: I1204 14:46:10.501831 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="3bde09932246f62233cddfa1e31a9c959096be21967b6b67dfbfba81b6e806e4" exitCode=0 Dec 04 14:46:10 crc kubenswrapper[4715]: I1204 14:46:10.501918 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"3bde09932246f62233cddfa1e31a9c959096be21967b6b67dfbfba81b6e806e4"} Dec 04 14:46:10 crc kubenswrapper[4715]: I1204 14:46:10.502183 4715 scope.go:117] "RemoveContainer" containerID="d90909ecca92cd565aba6796a5091796864a502b80565678d1de94a447496a1a" Dec 04 14:46:15 crc kubenswrapper[4715]: E1204 14:46:15.155873 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 04 14:46:15 crc kubenswrapper[4715]: E1204 14:46:15.156611 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-452mv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1314b612-6502-4cef-ade7-f05971b781a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:46:15 crc kubenswrapper[4715]: E1204 14:46:15.158456 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1314b612-6502-4cef-ade7-f05971b781a2" Dec 04 14:46:15 crc kubenswrapper[4715]: I1204 14:46:15.548213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d"} Dec 04 14:46:15 crc kubenswrapper[4715]: E1204 14:46:15.550157 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1314b612-6502-4cef-ade7-f05971b781a2" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.270735 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nndvj"] Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.273374 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.289270 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nndvj"] Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.425194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-catalog-content\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.425257 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-utilities\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.425398 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpbbk\" (UniqueName: \"kubernetes.io/projected/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-kube-api-access-qpbbk\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.527171 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-catalog-content\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.527242 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-utilities\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.527330 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpbbk\" (UniqueName: \"kubernetes.io/projected/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-kube-api-access-qpbbk\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.527668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-catalog-content\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.527730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-utilities\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.552218 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpbbk\" (UniqueName: \"kubernetes.io/projected/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-kube-api-access-qpbbk\") pod \"certified-operators-nndvj\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.592014 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:17 crc kubenswrapper[4715]: I1204 14:46:17.957706 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nndvj"] Dec 04 14:46:18 crc kubenswrapper[4715]: I1204 14:46:18.573704 4715 generic.go:334] "Generic (PLEG): container finished" podID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerID="415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b" exitCode=0 Dec 04 14:46:18 crc kubenswrapper[4715]: I1204 14:46:18.573784 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerDied","Data":"415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b"} Dec 04 14:46:18 crc kubenswrapper[4715]: I1204 14:46:18.574013 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerStarted","Data":"16a6a9cd16e1b0795a4b6ed2618784d9aae0785661a995edc42ef071abbe1054"} Dec 04 14:46:19 crc kubenswrapper[4715]: I1204 14:46:19.599795 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerStarted","Data":"ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c"} Dec 04 14:46:20 crc kubenswrapper[4715]: I1204 14:46:20.609705 4715 generic.go:334] "Generic (PLEG): container finished" podID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerID="ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c" exitCode=0 Dec 04 14:46:20 crc kubenswrapper[4715]: I1204 14:46:20.609760 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerDied","Data":"ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c"} Dec 04 14:46:21 crc kubenswrapper[4715]: I1204 14:46:21.619744 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerStarted","Data":"cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075"} Dec 04 14:46:21 crc kubenswrapper[4715]: I1204 14:46:21.638017 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nndvj" podStartSLOduration=1.902197954 podStartE2EDuration="4.637998991s" podCreationTimestamp="2025-12-04 14:46:17 +0000 UTC" firstStartedPulling="2025-12-04 14:46:18.575463063 +0000 UTC m=+2935.644181288" lastFinishedPulling="2025-12-04 14:46:21.31126411 +0000 UTC m=+2938.379982325" observedRunningTime="2025-12-04 14:46:21.636480679 +0000 UTC m=+2938.705198894" watchObservedRunningTime="2025-12-04 14:46:21.637998991 +0000 UTC m=+2938.706717206" Dec 04 14:46:27 crc kubenswrapper[4715]: I1204 14:46:27.593274 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:27 crc kubenswrapper[4715]: I1204 14:46:27.593941 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:27 crc kubenswrapper[4715]: I1204 14:46:27.656895 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:27 crc kubenswrapper[4715]: I1204 14:46:27.714416 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:30 crc kubenswrapper[4715]: I1204 14:46:30.666872 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nndvj"] Dec 04 14:46:30 crc kubenswrapper[4715]: I1204 14:46:30.667624 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nndvj" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="registry-server" containerID="cri-o://cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075" gracePeriod=2 Dec 04 14:46:30 crc kubenswrapper[4715]: I1204 14:46:30.717353 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.520266 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.638815 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpbbk\" (UniqueName: \"kubernetes.io/projected/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-kube-api-access-qpbbk\") pod \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.638865 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-catalog-content\") pod \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.638979 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-utilities\") pod \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\" (UID: \"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0\") " Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.640637 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-utilities" (OuterVolumeSpecName: "utilities") pod "522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" (UID: "522097d8-9a2c-46a9-9f00-a7cc4ded0fe0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.644571 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-kube-api-access-qpbbk" (OuterVolumeSpecName: "kube-api-access-qpbbk") pod "522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" (UID: "522097d8-9a2c-46a9-9f00-a7cc4ded0fe0"). InnerVolumeSpecName "kube-api-access-qpbbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.692169 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" (UID: "522097d8-9a2c-46a9-9f00-a7cc4ded0fe0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.706223 4715 generic.go:334] "Generic (PLEG): container finished" podID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerID="cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075" exitCode=0 Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.706275 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerDied","Data":"cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075"} Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.706343 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nndvj" event={"ID":"522097d8-9a2c-46a9-9f00-a7cc4ded0fe0","Type":"ContainerDied","Data":"16a6a9cd16e1b0795a4b6ed2618784d9aae0785661a995edc42ef071abbe1054"} Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.706338 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nndvj" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.706368 4715 scope.go:117] "RemoveContainer" containerID="cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.734300 4715 scope.go:117] "RemoveContainer" containerID="ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.741404 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpbbk\" (UniqueName: \"kubernetes.io/projected/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-kube-api-access-qpbbk\") on node \"crc\" DevicePath \"\"" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.741440 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.741449 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.747439 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nndvj"] Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.757073 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nndvj"] Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.783350 4715 scope.go:117] "RemoveContainer" containerID="415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.856251 4715 scope.go:117] "RemoveContainer" containerID="cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075" Dec 04 14:46:31 crc kubenswrapper[4715]: E1204 14:46:31.856666 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075\": container with ID starting with cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075 not found: ID does not exist" containerID="cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.856708 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075"} err="failed to get container status \"cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075\": rpc error: code = NotFound desc = could not find container \"cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075\": container with ID starting with cf15fa3e23dec1da02abd36b5beb37a95d52a650d265b81a9336932e8fca4075 not found: ID does not exist" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.856736 4715 scope.go:117] "RemoveContainer" containerID="ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c" Dec 04 14:46:31 crc kubenswrapper[4715]: E1204 14:46:31.857103 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c\": container with ID starting with ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c not found: ID does not exist" containerID="ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.857157 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c"} err="failed to get container status \"ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c\": rpc error: code = NotFound desc = could not find container \"ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c\": container with ID starting with ffb8e6c6f8d367f752f24d76d6ab9730c99c610c2d1d0a63b9c31fff4572fa7c not found: ID does not exist" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.857192 4715 scope.go:117] "RemoveContainer" containerID="415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b" Dec 04 14:46:31 crc kubenswrapper[4715]: E1204 14:46:31.857541 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b\": container with ID starting with 415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b not found: ID does not exist" containerID="415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b" Dec 04 14:46:31 crc kubenswrapper[4715]: I1204 14:46:31.857568 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b"} err="failed to get container status \"415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b\": rpc error: code = NotFound desc = could not find container \"415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b\": container with ID starting with 415ab410ec9906d01b95358b80ec360e17109c5e61f160d77298f63be734802b not found: ID does not exist" Dec 04 14:46:32 crc kubenswrapper[4715]: I1204 14:46:32.717888 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1314b612-6502-4cef-ade7-f05971b781a2","Type":"ContainerStarted","Data":"314d8e70d6e959e33152ab7f5920be5f7bde411be162420e97611af1cb027459"} Dec 04 14:46:32 crc kubenswrapper[4715]: I1204 14:46:32.747449 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.329799791 podStartE2EDuration="57.747425055s" podCreationTimestamp="2025-12-04 14:45:35 +0000 UTC" firstStartedPulling="2025-12-04 14:45:37.297087531 +0000 UTC m=+2894.365805746" lastFinishedPulling="2025-12-04 14:46:30.714712795 +0000 UTC m=+2947.783431010" observedRunningTime="2025-12-04 14:46:32.737609336 +0000 UTC m=+2949.806327561" watchObservedRunningTime="2025-12-04 14:46:32.747425055 +0000 UTC m=+2949.816143270" Dec 04 14:46:33 crc kubenswrapper[4715]: I1204 14:46:33.192792 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" path="/var/lib/kubelet/pods/522097d8-9a2c-46a9-9f00-a7cc4ded0fe0/volumes" Dec 04 14:48:38 crc kubenswrapper[4715]: I1204 14:48:38.758167 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:48:38 crc kubenswrapper[4715]: I1204 14:48:38.758727 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:49:08 crc kubenswrapper[4715]: I1204 14:49:08.758930 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:49:08 crc kubenswrapper[4715]: I1204 14:49:08.759457 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:49:38 crc kubenswrapper[4715]: I1204 14:49:38.759134 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:49:38 crc kubenswrapper[4715]: I1204 14:49:38.759766 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:49:38 crc kubenswrapper[4715]: I1204 14:49:38.759829 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:49:38 crc kubenswrapper[4715]: I1204 14:49:38.760638 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:49:38 crc kubenswrapper[4715]: I1204 14:49:38.760703 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" gracePeriod=600 Dec 04 14:49:38 crc kubenswrapper[4715]: E1204 14:49:38.886085 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:49:39 crc kubenswrapper[4715]: I1204 14:49:39.689507 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" exitCode=0 Dec 04 14:49:39 crc kubenswrapper[4715]: I1204 14:49:39.689575 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d"} Dec 04 14:49:39 crc kubenswrapper[4715]: I1204 14:49:39.689643 4715 scope.go:117] "RemoveContainer" containerID="3bde09932246f62233cddfa1e31a9c959096be21967b6b67dfbfba81b6e806e4" Dec 04 14:49:39 crc kubenswrapper[4715]: I1204 14:49:39.690324 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:49:39 crc kubenswrapper[4715]: E1204 14:49:39.690650 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:49:52 crc kubenswrapper[4715]: I1204 14:49:52.181355 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:49:52 crc kubenswrapper[4715]: E1204 14:49:52.182728 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:50:07 crc kubenswrapper[4715]: I1204 14:50:07.180561 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:50:07 crc kubenswrapper[4715]: E1204 14:50:07.181443 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:50:21 crc kubenswrapper[4715]: I1204 14:50:21.180548 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:50:21 crc kubenswrapper[4715]: E1204 14:50:21.181370 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:50:35 crc kubenswrapper[4715]: I1204 14:50:35.183423 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:50:35 crc kubenswrapper[4715]: E1204 14:50:35.184767 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:50:50 crc kubenswrapper[4715]: I1204 14:50:50.181818 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:50:50 crc kubenswrapper[4715]: E1204 14:50:50.182553 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:51:04 crc kubenswrapper[4715]: I1204 14:51:04.181125 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:51:04 crc kubenswrapper[4715]: E1204 14:51:04.182423 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:51:18 crc kubenswrapper[4715]: I1204 14:51:18.181420 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:51:18 crc kubenswrapper[4715]: E1204 14:51:18.182232 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:51:33 crc kubenswrapper[4715]: I1204 14:51:33.186709 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:51:33 crc kubenswrapper[4715]: E1204 14:51:33.187432 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:51:45 crc kubenswrapper[4715]: I1204 14:51:45.180984 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:51:45 crc kubenswrapper[4715]: E1204 14:51:45.181966 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:51:57 crc kubenswrapper[4715]: I1204 14:51:57.181096 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:51:57 crc kubenswrapper[4715]: E1204 14:51:57.182551 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:52:09 crc kubenswrapper[4715]: I1204 14:52:09.180014 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:52:09 crc kubenswrapper[4715]: E1204 14:52:09.180739 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:52:20 crc kubenswrapper[4715]: I1204 14:52:20.225375 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:52:20 crc kubenswrapper[4715]: E1204 14:52:20.226427 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:52:31 crc kubenswrapper[4715]: I1204 14:52:31.181450 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:52:31 crc kubenswrapper[4715]: E1204 14:52:31.182094 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:52:45 crc kubenswrapper[4715]: I1204 14:52:45.181148 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:52:45 crc kubenswrapper[4715]: E1204 14:52:45.181908 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:52:59 crc kubenswrapper[4715]: I1204 14:52:59.181070 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:52:59 crc kubenswrapper[4715]: E1204 14:52:59.182092 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:53:11 crc kubenswrapper[4715]: I1204 14:53:11.181173 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:53:11 crc kubenswrapper[4715]: E1204 14:53:11.182160 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:53:23 crc kubenswrapper[4715]: I1204 14:53:23.200423 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:53:23 crc kubenswrapper[4715]: E1204 14:53:23.201193 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:53:36 crc kubenswrapper[4715]: I1204 14:53:36.180609 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:53:36 crc kubenswrapper[4715]: E1204 14:53:36.181495 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:53:49 crc kubenswrapper[4715]: I1204 14:53:49.181191 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:53:49 crc kubenswrapper[4715]: E1204 14:53:49.181983 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:54:00 crc kubenswrapper[4715]: I1204 14:54:00.180568 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:54:00 crc kubenswrapper[4715]: E1204 14:54:00.181338 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:54:15 crc kubenswrapper[4715]: I1204 14:54:15.180891 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:54:15 crc kubenswrapper[4715]: E1204 14:54:15.181656 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:54:27 crc kubenswrapper[4715]: I1204 14:54:27.180764 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:54:27 crc kubenswrapper[4715]: E1204 14:54:27.181642 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 14:54:39 crc kubenswrapper[4715]: I1204 14:54:39.180743 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:54:39 crc kubenswrapper[4715]: I1204 14:54:39.655629 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"e2e7792c7b342edbef8b148ec4e4a9cd95546575483ca8dead24c87263498a7c"} Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.340148 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b9kpz"] Dec 04 14:54:57 crc kubenswrapper[4715]: E1204 14:54:57.341089 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="extract-utilities" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.341103 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="extract-utilities" Dec 04 14:54:57 crc kubenswrapper[4715]: E1204 14:54:57.341122 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="extract-content" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.341129 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="extract-content" Dec 04 14:54:57 crc kubenswrapper[4715]: E1204 14:54:57.341146 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="registry-server" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.341153 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="registry-server" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.341395 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="522097d8-9a2c-46a9-9f00-a7cc4ded0fe0" containerName="registry-server" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.343064 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.352305 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b9kpz"] Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.487368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e8b8414-517f-4857-ae4a-c4e131ef8822-catalog-content\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.487481 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9z8p\" (UniqueName: \"kubernetes.io/projected/5e8b8414-517f-4857-ae4a-c4e131ef8822-kube-api-access-p9z8p\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.487531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e8b8414-517f-4857-ae4a-c4e131ef8822-utilities\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.589571 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9z8p\" (UniqueName: \"kubernetes.io/projected/5e8b8414-517f-4857-ae4a-c4e131ef8822-kube-api-access-p9z8p\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.589652 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e8b8414-517f-4857-ae4a-c4e131ef8822-utilities\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.589711 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e8b8414-517f-4857-ae4a-c4e131ef8822-catalog-content\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.590193 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e8b8414-517f-4857-ae4a-c4e131ef8822-catalog-content\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.590495 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e8b8414-517f-4857-ae4a-c4e131ef8822-utilities\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.626357 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9z8p\" (UniqueName: \"kubernetes.io/projected/5e8b8414-517f-4857-ae4a-c4e131ef8822-kube-api-access-p9z8p\") pod \"redhat-operators-b9kpz\" (UID: \"5e8b8414-517f-4857-ae4a-c4e131ef8822\") " pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:57 crc kubenswrapper[4715]: I1204 14:54:57.675311 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:54:58 crc kubenswrapper[4715]: I1204 14:54:58.170141 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b9kpz"] Dec 04 14:54:58 crc kubenswrapper[4715]: I1204 14:54:58.898940 4715 generic.go:334] "Generic (PLEG): container finished" podID="5e8b8414-517f-4857-ae4a-c4e131ef8822" containerID="0a0983360fa7d8587ed4c8e67e5d3a51882ee4fe29778ee9d95fb103644c8883" exitCode=0 Dec 04 14:54:58 crc kubenswrapper[4715]: I1204 14:54:58.899010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b9kpz" event={"ID":"5e8b8414-517f-4857-ae4a-c4e131ef8822","Type":"ContainerDied","Data":"0a0983360fa7d8587ed4c8e67e5d3a51882ee4fe29778ee9d95fb103644c8883"} Dec 04 14:54:58 crc kubenswrapper[4715]: I1204 14:54:58.899288 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b9kpz" event={"ID":"5e8b8414-517f-4857-ae4a-c4e131ef8822","Type":"ContainerStarted","Data":"d8aff8f4fe6529f4e273f87bac6e86a4dd0047c47b59d3aee89599a5c3c92200"} Dec 04 14:54:58 crc kubenswrapper[4715]: I1204 14:54:58.901323 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:55:08 crc kubenswrapper[4715]: I1204 14:55:08.996766 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b9kpz" event={"ID":"5e8b8414-517f-4857-ae4a-c4e131ef8822","Type":"ContainerStarted","Data":"7d9ad4e41c8e57dd7ac2d5870ccff5de29ac283340e9ce1c58f028f9a7fdd6a4"} Dec 04 14:55:12 crc kubenswrapper[4715]: I1204 14:55:12.023165 4715 generic.go:334] "Generic (PLEG): container finished" podID="5e8b8414-517f-4857-ae4a-c4e131ef8822" containerID="7d9ad4e41c8e57dd7ac2d5870ccff5de29ac283340e9ce1c58f028f9a7fdd6a4" exitCode=0 Dec 04 14:55:12 crc kubenswrapper[4715]: I1204 14:55:12.023243 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b9kpz" event={"ID":"5e8b8414-517f-4857-ae4a-c4e131ef8822","Type":"ContainerDied","Data":"7d9ad4e41c8e57dd7ac2d5870ccff5de29ac283340e9ce1c58f028f9a7fdd6a4"} Dec 04 14:55:14 crc kubenswrapper[4715]: I1204 14:55:14.044486 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b9kpz" event={"ID":"5e8b8414-517f-4857-ae4a-c4e131ef8822","Type":"ContainerStarted","Data":"dcc281463a3a25cf9bc467d091f6f81930e9731be3d5a2ed51e2889fafcea439"} Dec 04 14:55:14 crc kubenswrapper[4715]: I1204 14:55:14.076522 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b9kpz" podStartSLOduration=2.683090675 podStartE2EDuration="17.076496652s" podCreationTimestamp="2025-12-04 14:54:57 +0000 UTC" firstStartedPulling="2025-12-04 14:54:58.901014997 +0000 UTC m=+3455.969733212" lastFinishedPulling="2025-12-04 14:55:13.294420984 +0000 UTC m=+3470.363139189" observedRunningTime="2025-12-04 14:55:14.061644136 +0000 UTC m=+3471.130362371" watchObservedRunningTime="2025-12-04 14:55:14.076496652 +0000 UTC m=+3471.145214887" Dec 04 14:55:17 crc kubenswrapper[4715]: I1204 14:55:17.675733 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:55:17 crc kubenswrapper[4715]: I1204 14:55:17.676336 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:55:18 crc kubenswrapper[4715]: I1204 14:55:18.804754 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b9kpz" podUID="5e8b8414-517f-4857-ae4a-c4e131ef8822" containerName="registry-server" probeResult="failure" output=< Dec 04 14:55:18 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 04 14:55:18 crc kubenswrapper[4715]: > Dec 04 14:55:27 crc kubenswrapper[4715]: I1204 14:55:27.721611 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:55:27 crc kubenswrapper[4715]: I1204 14:55:27.773804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b9kpz" Dec 04 14:55:28 crc kubenswrapper[4715]: I1204 14:55:28.361096 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b9kpz"] Dec 04 14:55:28 crc kubenswrapper[4715]: I1204 14:55:28.545202 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lsst9"] Dec 04 14:55:28 crc kubenswrapper[4715]: I1204 14:55:28.545817 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lsst9" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="registry-server" containerID="cri-o://ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102" gracePeriod=2 Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.074746 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.201128 4715 generic.go:334] "Generic (PLEG): container finished" podID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerID="ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102" exitCode=0 Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.201285 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lsst9" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.203921 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lsst9" event={"ID":"69d2a077-7d27-4c2f-82c5-d14628dbe954","Type":"ContainerDied","Data":"ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102"} Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.203972 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lsst9" event={"ID":"69d2a077-7d27-4c2f-82c5-d14628dbe954","Type":"ContainerDied","Data":"c2eb5e3b431281d045080e8e06976475e449ed7fc032d27e1fa6252e56ea177b"} Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.204008 4715 scope.go:117] "RemoveContainer" containerID="ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.229779 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-catalog-content\") pod \"69d2a077-7d27-4c2f-82c5-d14628dbe954\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.229845 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd5md\" (UniqueName: \"kubernetes.io/projected/69d2a077-7d27-4c2f-82c5-d14628dbe954-kube-api-access-sd5md\") pod \"69d2a077-7d27-4c2f-82c5-d14628dbe954\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.229897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-utilities\") pod \"69d2a077-7d27-4c2f-82c5-d14628dbe954\" (UID: \"69d2a077-7d27-4c2f-82c5-d14628dbe954\") " Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.231300 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-utilities" (OuterVolumeSpecName: "utilities") pod "69d2a077-7d27-4c2f-82c5-d14628dbe954" (UID: "69d2a077-7d27-4c2f-82c5-d14628dbe954"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.231835 4715 scope.go:117] "RemoveContainer" containerID="d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.237083 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69d2a077-7d27-4c2f-82c5-d14628dbe954-kube-api-access-sd5md" (OuterVolumeSpecName: "kube-api-access-sd5md") pod "69d2a077-7d27-4c2f-82c5-d14628dbe954" (UID: "69d2a077-7d27-4c2f-82c5-d14628dbe954"). InnerVolumeSpecName "kube-api-access-sd5md". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.299733 4715 scope.go:117] "RemoveContainer" containerID="94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.330650 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69d2a077-7d27-4c2f-82c5-d14628dbe954" (UID: "69d2a077-7d27-4c2f-82c5-d14628dbe954"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.332350 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.332482 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd5md\" (UniqueName: \"kubernetes.io/projected/69d2a077-7d27-4c2f-82c5-d14628dbe954-kube-api-access-sd5md\") on node \"crc\" DevicePath \"\"" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.332549 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d2a077-7d27-4c2f-82c5-d14628dbe954-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.342866 4715 scope.go:117] "RemoveContainer" containerID="ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102" Dec 04 14:55:29 crc kubenswrapper[4715]: E1204 14:55:29.343570 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102\": container with ID starting with ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102 not found: ID does not exist" containerID="ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.343605 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102"} err="failed to get container status \"ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102\": rpc error: code = NotFound desc = could not find container \"ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102\": container with ID starting with ea67a95bbb54a740e09d42a3947ab61b8026d814542d939141783a800b73b102 not found: ID does not exist" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.343625 4715 scope.go:117] "RemoveContainer" containerID="d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b" Dec 04 14:55:29 crc kubenswrapper[4715]: E1204 14:55:29.344106 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b\": container with ID starting with d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b not found: ID does not exist" containerID="d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.344233 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b"} err="failed to get container status \"d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b\": rpc error: code = NotFound desc = could not find container \"d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b\": container with ID starting with d7aedf73e215ceab6261f7ac67be796b4fbf275c1b4a29e11987553775c2fa7b not found: ID does not exist" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.344338 4715 scope.go:117] "RemoveContainer" containerID="94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1" Dec 04 14:55:29 crc kubenswrapper[4715]: E1204 14:55:29.344813 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1\": container with ID starting with 94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1 not found: ID does not exist" containerID="94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.344837 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1"} err="failed to get container status \"94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1\": rpc error: code = NotFound desc = could not find container \"94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1\": container with ID starting with 94d24470b9c58e0e9944997778cd676eb68351d05205bccafc9a1baf083beca1 not found: ID does not exist" Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.534022 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lsst9"] Dec 04 14:55:29 crc kubenswrapper[4715]: I1204 14:55:29.543341 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lsst9"] Dec 04 14:55:31 crc kubenswrapper[4715]: I1204 14:55:31.195948 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" path="/var/lib/kubelet/pods/69d2a077-7d27-4c2f-82c5-d14628dbe954/volumes" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.483663 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vjkl9"] Dec 04 14:56:01 crc kubenswrapper[4715]: E1204 14:56:01.484589 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="registry-server" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.484602 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="registry-server" Dec 04 14:56:01 crc kubenswrapper[4715]: E1204 14:56:01.484622 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="extract-utilities" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.484628 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="extract-utilities" Dec 04 14:56:01 crc kubenswrapper[4715]: E1204 14:56:01.484642 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="extract-content" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.484647 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="extract-content" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.484853 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="69d2a077-7d27-4c2f-82c5-d14628dbe954" containerName="registry-server" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.486271 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.493939 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vjkl9"] Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.573927 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-utilities\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.574254 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-catalog-content\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.574980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/daa28e83-ddfa-4d13-85dc-43389fa12ef2-kube-api-access-kghvc\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.676639 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/daa28e83-ddfa-4d13-85dc-43389fa12ef2-kube-api-access-kghvc\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.676710 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-utilities\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.676744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-catalog-content\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.677309 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-utilities\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.677354 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-catalog-content\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.698147 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/daa28e83-ddfa-4d13-85dc-43389fa12ef2-kube-api-access-kghvc\") pod \"community-operators-vjkl9\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:01 crc kubenswrapper[4715]: I1204 14:56:01.813356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:02 crc kubenswrapper[4715]: I1204 14:56:02.330756 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vjkl9"] Dec 04 14:56:02 crc kubenswrapper[4715]: I1204 14:56:02.517355 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerStarted","Data":"7f400d5f5fb6fd6e4f776cd7b6f114bfa41fe010ce28ef5ed965735c20bd2804"} Dec 04 14:56:03 crc kubenswrapper[4715]: I1204 14:56:03.531194 4715 generic.go:334] "Generic (PLEG): container finished" podID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerID="d40aca24e416885673aa8f7ef3fb2ec764f6d4715303f4f0707b273c1cdd7c54" exitCode=0 Dec 04 14:56:03 crc kubenswrapper[4715]: I1204 14:56:03.531268 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerDied","Data":"d40aca24e416885673aa8f7ef3fb2ec764f6d4715303f4f0707b273c1cdd7c54"} Dec 04 14:56:05 crc kubenswrapper[4715]: I1204 14:56:05.565812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerStarted","Data":"1ce732253d8801cea9a33a0a4ff55e1f2cdd60c5d98416e37805d3d3b40ec46a"} Dec 04 14:56:06 crc kubenswrapper[4715]: I1204 14:56:06.577080 4715 generic.go:334] "Generic (PLEG): container finished" podID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerID="1ce732253d8801cea9a33a0a4ff55e1f2cdd60c5d98416e37805d3d3b40ec46a" exitCode=0 Dec 04 14:56:06 crc kubenswrapper[4715]: I1204 14:56:06.577135 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerDied","Data":"1ce732253d8801cea9a33a0a4ff55e1f2cdd60c5d98416e37805d3d3b40ec46a"} Dec 04 14:56:07 crc kubenswrapper[4715]: I1204 14:56:07.588757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerStarted","Data":"e71838da48d31258d6653e5bf2b0783c1fc9018c0afcff8674561fd6c1a6bbf7"} Dec 04 14:56:07 crc kubenswrapper[4715]: I1204 14:56:07.614695 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vjkl9" podStartSLOduration=2.9182511509999998 podStartE2EDuration="6.61466822s" podCreationTimestamp="2025-12-04 14:56:01 +0000 UTC" firstStartedPulling="2025-12-04 14:56:03.534472495 +0000 UTC m=+3520.603190750" lastFinishedPulling="2025-12-04 14:56:07.230889594 +0000 UTC m=+3524.299607819" observedRunningTime="2025-12-04 14:56:07.603784903 +0000 UTC m=+3524.672503118" watchObservedRunningTime="2025-12-04 14:56:07.61466822 +0000 UTC m=+3524.683386435" Dec 04 14:56:11 crc kubenswrapper[4715]: I1204 14:56:11.813548 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:11 crc kubenswrapper[4715]: I1204 14:56:11.814240 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:11 crc kubenswrapper[4715]: I1204 14:56:11.893923 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:12 crc kubenswrapper[4715]: I1204 14:56:12.686368 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:12 crc kubenswrapper[4715]: I1204 14:56:12.737720 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vjkl9"] Dec 04 14:56:14 crc kubenswrapper[4715]: I1204 14:56:14.665863 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vjkl9" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="registry-server" containerID="cri-o://e71838da48d31258d6653e5bf2b0783c1fc9018c0afcff8674561fd6c1a6bbf7" gracePeriod=2 Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.676698 4715 generic.go:334] "Generic (PLEG): container finished" podID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerID="e71838da48d31258d6653e5bf2b0783c1fc9018c0afcff8674561fd6c1a6bbf7" exitCode=0 Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.676777 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerDied","Data":"e71838da48d31258d6653e5bf2b0783c1fc9018c0afcff8674561fd6c1a6bbf7"} Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.676963 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjkl9" event={"ID":"daa28e83-ddfa-4d13-85dc-43389fa12ef2","Type":"ContainerDied","Data":"7f400d5f5fb6fd6e4f776cd7b6f114bfa41fe010ce28ef5ed965735c20bd2804"} Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.676980 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f400d5f5fb6fd6e4f776cd7b6f114bfa41fe010ce28ef5ed965735c20bd2804" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.700768 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.872226 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-catalog-content\") pod \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.872333 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-utilities\") pod \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.872484 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/daa28e83-ddfa-4d13-85dc-43389fa12ef2-kube-api-access-kghvc\") pod \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\" (UID: \"daa28e83-ddfa-4d13-85dc-43389fa12ef2\") " Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.873311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-utilities" (OuterVolumeSpecName: "utilities") pod "daa28e83-ddfa-4d13-85dc-43389fa12ef2" (UID: "daa28e83-ddfa-4d13-85dc-43389fa12ef2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.885361 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa28e83-ddfa-4d13-85dc-43389fa12ef2-kube-api-access-kghvc" (OuterVolumeSpecName: "kube-api-access-kghvc") pod "daa28e83-ddfa-4d13-85dc-43389fa12ef2" (UID: "daa28e83-ddfa-4d13-85dc-43389fa12ef2"). InnerVolumeSpecName "kube-api-access-kghvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.925313 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "daa28e83-ddfa-4d13-85dc-43389fa12ef2" (UID: "daa28e83-ddfa-4d13-85dc-43389fa12ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.974329 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.974373 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/daa28e83-ddfa-4d13-85dc-43389fa12ef2-kube-api-access-kghvc\") on node \"crc\" DevicePath \"\"" Dec 04 14:56:15 crc kubenswrapper[4715]: I1204 14:56:15.974389 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daa28e83-ddfa-4d13-85dc-43389fa12ef2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:56:16 crc kubenswrapper[4715]: I1204 14:56:16.689288 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjkl9" Dec 04 14:56:16 crc kubenswrapper[4715]: I1204 14:56:16.742893 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vjkl9"] Dec 04 14:56:16 crc kubenswrapper[4715]: I1204 14:56:16.755909 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vjkl9"] Dec 04 14:56:17 crc kubenswrapper[4715]: I1204 14:56:17.192327 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" path="/var/lib/kubelet/pods/daa28e83-ddfa-4d13-85dc-43389fa12ef2/volumes" Dec 04 14:57:08 crc kubenswrapper[4715]: I1204 14:57:08.758755 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:57:08 crc kubenswrapper[4715]: I1204 14:57:08.759367 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:57:38 crc kubenswrapper[4715]: I1204 14:57:38.758270 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:57:38 crc kubenswrapper[4715]: I1204 14:57:38.759085 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:57:57 crc kubenswrapper[4715]: I1204 14:57:57.684677 4715 generic.go:334] "Generic (PLEG): container finished" podID="1314b612-6502-4cef-ade7-f05971b781a2" containerID="314d8e70d6e959e33152ab7f5920be5f7bde411be162420e97611af1cb027459" exitCode=0 Dec 04 14:57:57 crc kubenswrapper[4715]: I1204 14:57:57.684785 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1314b612-6502-4cef-ade7-f05971b781a2","Type":"ContainerDied","Data":"314d8e70d6e959e33152ab7f5920be5f7bde411be162420e97611af1cb027459"} Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.090836 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.214866 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-workdir\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.214932 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config-secret\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.214987 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-452mv\" (UniqueName: \"kubernetes.io/projected/1314b612-6502-4cef-ade7-f05971b781a2-kube-api-access-452mv\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.215157 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.215191 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ssh-key\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.215249 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-config-data\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.215314 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-temporary\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.215370 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ca-certs\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.215479 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config\") pod \"1314b612-6502-4cef-ade7-f05971b781a2\" (UID: \"1314b612-6502-4cef-ade7-f05971b781a2\") " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.216764 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.216782 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-config-data" (OuterVolumeSpecName: "config-data") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.219996 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.220394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.220596 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1314b612-6502-4cef-ade7-f05971b781a2-kube-api-access-452mv" (OuterVolumeSpecName: "kube-api-access-452mv") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "kube-api-access-452mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.247094 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.247975 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.248823 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.268723 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1314b612-6502-4cef-ade7-f05971b781a2" (UID: "1314b612-6502-4cef-ade7-f05971b781a2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.317998 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318058 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318073 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318086 4715 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318098 4715 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318108 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318122 4715 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1314b612-6502-4cef-ade7-f05971b781a2-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318132 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1314b612-6502-4cef-ade7-f05971b781a2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.318143 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-452mv\" (UniqueName: \"kubernetes.io/projected/1314b612-6502-4cef-ade7-f05971b781a2-kube-api-access-452mv\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.337844 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.419327 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.709753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1314b612-6502-4cef-ade7-f05971b781a2","Type":"ContainerDied","Data":"dc5e43ecc47a9f01960a8abb1e3c5a853eea68cd44aac2d3ae8577d8bb3415a7"} Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.709806 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc5e43ecc47a9f01960a8abb1e3c5a853eea68cd44aac2d3ae8577d8bb3415a7" Dec 04 14:57:59 crc kubenswrapper[4715]: I1204 14:57:59.709856 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 14:58:08 crc kubenswrapper[4715]: I1204 14:58:08.758278 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:58:08 crc kubenswrapper[4715]: I1204 14:58:08.758804 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:58:08 crc kubenswrapper[4715]: I1204 14:58:08.758854 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 14:58:08 crc kubenswrapper[4715]: I1204 14:58:08.759552 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e2e7792c7b342edbef8b148ec4e4a9cd95546575483ca8dead24c87263498a7c"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:58:08 crc kubenswrapper[4715]: I1204 14:58:08.759604 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://e2e7792c7b342edbef8b148ec4e4a9cd95546575483ca8dead24c87263498a7c" gracePeriod=600 Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.015544 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 14:58:09 crc kubenswrapper[4715]: E1204 14:58:09.016312 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1314b612-6502-4cef-ade7-f05971b781a2" containerName="tempest-tests-tempest-tests-runner" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.016330 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1314b612-6502-4cef-ade7-f05971b781a2" containerName="tempest-tests-tempest-tests-runner" Dec 04 14:58:09 crc kubenswrapper[4715]: E1204 14:58:09.016347 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="extract-content" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.016354 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="extract-content" Dec 04 14:58:09 crc kubenswrapper[4715]: E1204 14:58:09.016372 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="registry-server" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.016379 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="registry-server" Dec 04 14:58:09 crc kubenswrapper[4715]: E1204 14:58:09.016401 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="extract-utilities" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.016407 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="extract-utilities" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.016578 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1314b612-6502-4cef-ade7-f05971b781a2" containerName="tempest-tests-tempest-tests-runner" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.016591 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa28e83-ddfa-4d13-85dc-43389fa12ef2" containerName="registry-server" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.017517 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.023517 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.057487 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-h87fs" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.103615 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.104019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4jd6\" (UniqueName: \"kubernetes.io/projected/212e2726-336d-42d9-942a-36642b0c1c98-kube-api-access-x4jd6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.206490 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4jd6\" (UniqueName: \"kubernetes.io/projected/212e2726-336d-42d9-942a-36642b0c1c98-kube-api-access-x4jd6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.206835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.208985 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.251782 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4jd6\" (UniqueName: \"kubernetes.io/projected/212e2726-336d-42d9-942a-36642b0c1c98-kube-api-access-x4jd6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.347326 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"212e2726-336d-42d9-942a-36642b0c1c98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.397735 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.806937 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="e2e7792c7b342edbef8b148ec4e4a9cd95546575483ca8dead24c87263498a7c" exitCode=0 Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.807043 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"e2e7792c7b342edbef8b148ec4e4a9cd95546575483ca8dead24c87263498a7c"} Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.807358 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375"} Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.807393 4715 scope.go:117] "RemoveContainer" containerID="f94fe4bc2f1a766c1d4c6207fdc00cc93839a24a4f893a8aec7116585872804d" Dec 04 14:58:09 crc kubenswrapper[4715]: I1204 14:58:09.845090 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 14:58:09 crc kubenswrapper[4715]: W1204 14:58:09.853643 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod212e2726_336d_42d9_942a_36642b0c1c98.slice/crio-6b759aaa29dcbba2c0864b781960643d3589d99725f80eedf585774f4b649453 WatchSource:0}: Error finding container 6b759aaa29dcbba2c0864b781960643d3589d99725f80eedf585774f4b649453: Status 404 returned error can't find the container with id 6b759aaa29dcbba2c0864b781960643d3589d99725f80eedf585774f4b649453 Dec 04 14:58:10 crc kubenswrapper[4715]: I1204 14:58:10.819662 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"212e2726-336d-42d9-942a-36642b0c1c98","Type":"ContainerStarted","Data":"6b759aaa29dcbba2c0864b781960643d3589d99725f80eedf585774f4b649453"} Dec 04 14:58:11 crc kubenswrapper[4715]: I1204 14:58:11.835164 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"212e2726-336d-42d9-942a-36642b0c1c98","Type":"ContainerStarted","Data":"04570f1bf17ab05206b45d3ebe0202e9ad60e2ad86c75cceabbcc9274930864e"} Dec 04 14:58:11 crc kubenswrapper[4715]: I1204 14:58:11.854104 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.08048527 podStartE2EDuration="3.854084113s" podCreationTimestamp="2025-12-04 14:58:08 +0000 UTC" firstStartedPulling="2025-12-04 14:58:09.856896888 +0000 UTC m=+3646.925615103" lastFinishedPulling="2025-12-04 14:58:10.630495731 +0000 UTC m=+3647.699213946" observedRunningTime="2025-12-04 14:58:11.851952804 +0000 UTC m=+3648.920671019" watchObservedRunningTime="2025-12-04 14:58:11.854084113 +0000 UTC m=+3648.922802318" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.688296 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn5hf/must-gather-88prb"] Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.690612 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.692480 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zn5hf"/"openshift-service-ca.crt" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.692697 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zn5hf"/"default-dockercfg-qtwrn" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.693229 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zn5hf"/"kube-root-ca.crt" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.701112 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zn5hf/must-gather-88prb"] Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.801881 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbc7b\" (UniqueName: \"kubernetes.io/projected/ba37b5d6-3f67-4c36-bfd0-052e0989023a-kube-api-access-nbc7b\") pod \"must-gather-88prb\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.801967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ba37b5d6-3f67-4c36-bfd0-052e0989023a-must-gather-output\") pod \"must-gather-88prb\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.903909 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbc7b\" (UniqueName: \"kubernetes.io/projected/ba37b5d6-3f67-4c36-bfd0-052e0989023a-kube-api-access-nbc7b\") pod \"must-gather-88prb\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.904518 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ba37b5d6-3f67-4c36-bfd0-052e0989023a-must-gather-output\") pod \"must-gather-88prb\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.905065 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ba37b5d6-3f67-4c36-bfd0-052e0989023a-must-gather-output\") pod \"must-gather-88prb\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:33 crc kubenswrapper[4715]: I1204 14:58:33.930851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbc7b\" (UniqueName: \"kubernetes.io/projected/ba37b5d6-3f67-4c36-bfd0-052e0989023a-kube-api-access-nbc7b\") pod \"must-gather-88prb\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:34 crc kubenswrapper[4715]: I1204 14:58:34.005967 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 14:58:34 crc kubenswrapper[4715]: I1204 14:58:34.495785 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zn5hf/must-gather-88prb"] Dec 04 14:58:34 crc kubenswrapper[4715]: W1204 14:58:34.509517 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba37b5d6_3f67_4c36_bfd0_052e0989023a.slice/crio-6554a7aafa329c194c1bee9c157b9a4c16b20e52d623f7bf03f59b022deb8563 WatchSource:0}: Error finding container 6554a7aafa329c194c1bee9c157b9a4c16b20e52d623f7bf03f59b022deb8563: Status 404 returned error can't find the container with id 6554a7aafa329c194c1bee9c157b9a4c16b20e52d623f7bf03f59b022deb8563 Dec 04 14:58:35 crc kubenswrapper[4715]: I1204 14:58:35.073184 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/must-gather-88prb" event={"ID":"ba37b5d6-3f67-4c36-bfd0-052e0989023a","Type":"ContainerStarted","Data":"6554a7aafa329c194c1bee9c157b9a4c16b20e52d623f7bf03f59b022deb8563"} Dec 04 14:58:39 crc kubenswrapper[4715]: I1204 14:58:39.115127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/must-gather-88prb" event={"ID":"ba37b5d6-3f67-4c36-bfd0-052e0989023a","Type":"ContainerStarted","Data":"e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef"} Dec 04 14:58:39 crc kubenswrapper[4715]: I1204 14:58:39.115622 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/must-gather-88prb" event={"ID":"ba37b5d6-3f67-4c36-bfd0-052e0989023a","Type":"ContainerStarted","Data":"49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865"} Dec 04 14:58:39 crc kubenswrapper[4715]: I1204 14:58:39.138529 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zn5hf/must-gather-88prb" podStartSLOduration=2.475787039 podStartE2EDuration="6.138509104s" podCreationTimestamp="2025-12-04 14:58:33 +0000 UTC" firstStartedPulling="2025-12-04 14:58:34.511751074 +0000 UTC m=+3671.580469289" lastFinishedPulling="2025-12-04 14:58:38.174473139 +0000 UTC m=+3675.243191354" observedRunningTime="2025-12-04 14:58:39.1321859 +0000 UTC m=+3676.200904125" watchObservedRunningTime="2025-12-04 14:58:39.138509104 +0000 UTC m=+3676.207227319" Dec 04 14:58:41 crc kubenswrapper[4715]: E1204 14:58:41.247826 4715 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.198:45538->38.102.83.198:41743: write tcp 38.102.83.198:45538->38.102.83.198:41743: write: broken pipe Dec 04 14:58:41 crc kubenswrapper[4715]: I1204 14:58:41.861940 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-844vl"] Dec 04 14:58:41 crc kubenswrapper[4715]: I1204 14:58:41.863474 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:41 crc kubenswrapper[4715]: I1204 14:58:41.971144 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b8nv\" (UniqueName: \"kubernetes.io/projected/5ff445eb-30f3-4159-b738-274435a542b5-kube-api-access-9b8nv\") pod \"crc-debug-844vl\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:41 crc kubenswrapper[4715]: I1204 14:58:41.971193 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ff445eb-30f3-4159-b738-274435a542b5-host\") pod \"crc-debug-844vl\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:42 crc kubenswrapper[4715]: I1204 14:58:42.072968 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b8nv\" (UniqueName: \"kubernetes.io/projected/5ff445eb-30f3-4159-b738-274435a542b5-kube-api-access-9b8nv\") pod \"crc-debug-844vl\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:42 crc kubenswrapper[4715]: I1204 14:58:42.073106 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ff445eb-30f3-4159-b738-274435a542b5-host\") pod \"crc-debug-844vl\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:42 crc kubenswrapper[4715]: I1204 14:58:42.073283 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ff445eb-30f3-4159-b738-274435a542b5-host\") pod \"crc-debug-844vl\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:42 crc kubenswrapper[4715]: I1204 14:58:42.105389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b8nv\" (UniqueName: \"kubernetes.io/projected/5ff445eb-30f3-4159-b738-274435a542b5-kube-api-access-9b8nv\") pod \"crc-debug-844vl\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:42 crc kubenswrapper[4715]: I1204 14:58:42.183635 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:58:42 crc kubenswrapper[4715]: W1204 14:58:42.234838 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ff445eb_30f3_4159_b738_274435a542b5.slice/crio-e150c2b7eabc95d0e168b7ff595f81e19b562140c6f09512fa1e2be5e8d8b811 WatchSource:0}: Error finding container e150c2b7eabc95d0e168b7ff595f81e19b562140c6f09512fa1e2be5e8d8b811: Status 404 returned error can't find the container with id e150c2b7eabc95d0e168b7ff595f81e19b562140c6f09512fa1e2be5e8d8b811 Dec 04 14:58:43 crc kubenswrapper[4715]: I1204 14:58:43.157280 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-844vl" event={"ID":"5ff445eb-30f3-4159-b738-274435a542b5","Type":"ContainerStarted","Data":"e150c2b7eabc95d0e168b7ff595f81e19b562140c6f09512fa1e2be5e8d8b811"} Dec 04 14:58:57 crc kubenswrapper[4715]: I1204 14:58:57.305136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-844vl" event={"ID":"5ff445eb-30f3-4159-b738-274435a542b5","Type":"ContainerStarted","Data":"c8da63f77942de22144be60f5f2889044db8eaca19f6d424edba997f9d85a6c7"} Dec 04 14:58:57 crc kubenswrapper[4715]: I1204 14:58:57.325248 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zn5hf/crc-debug-844vl" podStartSLOduration=1.6819248070000001 podStartE2EDuration="16.325226049s" podCreationTimestamp="2025-12-04 14:58:41 +0000 UTC" firstStartedPulling="2025-12-04 14:58:42.238328643 +0000 UTC m=+3679.307046858" lastFinishedPulling="2025-12-04 14:58:56.881629885 +0000 UTC m=+3693.950348100" observedRunningTime="2025-12-04 14:58:57.321105316 +0000 UTC m=+3694.389823531" watchObservedRunningTime="2025-12-04 14:58:57.325226049 +0000 UTC m=+3694.393944274" Dec 04 14:59:44 crc kubenswrapper[4715]: E1204 14:59:44.870026 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ff445eb_30f3_4159_b738_274435a542b5.slice/crio-c8da63f77942de22144be60f5f2889044db8eaca19f6d424edba997f9d85a6c7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ff445eb_30f3_4159_b738_274435a542b5.slice/crio-conmon-c8da63f77942de22144be60f5f2889044db8eaca19f6d424edba997f9d85a6c7.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:59:44 crc kubenswrapper[4715]: I1204 14:59:44.996571 4715 generic.go:334] "Generic (PLEG): container finished" podID="5ff445eb-30f3-4159-b738-274435a542b5" containerID="c8da63f77942de22144be60f5f2889044db8eaca19f6d424edba997f9d85a6c7" exitCode=0 Dec 04 14:59:44 crc kubenswrapper[4715]: I1204 14:59:44.996643 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-844vl" event={"ID":"5ff445eb-30f3-4159-b738-274435a542b5","Type":"ContainerDied","Data":"c8da63f77942de22144be60f5f2889044db8eaca19f6d424edba997f9d85a6c7"} Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.116339 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.149227 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-844vl"] Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.159571 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-844vl"] Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.208722 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b8nv\" (UniqueName: \"kubernetes.io/projected/5ff445eb-30f3-4159-b738-274435a542b5-kube-api-access-9b8nv\") pod \"5ff445eb-30f3-4159-b738-274435a542b5\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.208806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ff445eb-30f3-4159-b738-274435a542b5-host\") pod \"5ff445eb-30f3-4159-b738-274435a542b5\" (UID: \"5ff445eb-30f3-4159-b738-274435a542b5\") " Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.208958 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ff445eb-30f3-4159-b738-274435a542b5-host" (OuterVolumeSpecName: "host") pod "5ff445eb-30f3-4159-b738-274435a542b5" (UID: "5ff445eb-30f3-4159-b738-274435a542b5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.209268 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ff445eb-30f3-4159-b738-274435a542b5-host\") on node \"crc\" DevicePath \"\"" Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.215110 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff445eb-30f3-4159-b738-274435a542b5-kube-api-access-9b8nv" (OuterVolumeSpecName: "kube-api-access-9b8nv") pod "5ff445eb-30f3-4159-b738-274435a542b5" (UID: "5ff445eb-30f3-4159-b738-274435a542b5"). InnerVolumeSpecName "kube-api-access-9b8nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:59:46 crc kubenswrapper[4715]: I1204 14:59:46.311596 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b8nv\" (UniqueName: \"kubernetes.io/projected/5ff445eb-30f3-4159-b738-274435a542b5-kube-api-access-9b8nv\") on node \"crc\" DevicePath \"\"" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.018943 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e150c2b7eabc95d0e168b7ff595f81e19b562140c6f09512fa1e2be5e8d8b811" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.019002 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-844vl" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.191005 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff445eb-30f3-4159-b738-274435a542b5" path="/var/lib/kubelet/pods/5ff445eb-30f3-4159-b738-274435a542b5/volumes" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.319081 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-56h9p"] Dec 04 14:59:47 crc kubenswrapper[4715]: E1204 14:59:47.319527 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff445eb-30f3-4159-b738-274435a542b5" containerName="container-00" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.319547 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff445eb-30f3-4159-b738-274435a542b5" containerName="container-00" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.319762 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff445eb-30f3-4159-b738-274435a542b5" containerName="container-00" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.320425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.429566 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1cc1035-1281-4e76-bb06-6de2949f4e25-host\") pod \"crc-debug-56h9p\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.429632 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxhtq\" (UniqueName: \"kubernetes.io/projected/d1cc1035-1281-4e76-bb06-6de2949f4e25-kube-api-access-gxhtq\") pod \"crc-debug-56h9p\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.530949 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1cc1035-1281-4e76-bb06-6de2949f4e25-host\") pod \"crc-debug-56h9p\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.531001 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxhtq\" (UniqueName: \"kubernetes.io/projected/d1cc1035-1281-4e76-bb06-6de2949f4e25-kube-api-access-gxhtq\") pod \"crc-debug-56h9p\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.531170 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1cc1035-1281-4e76-bb06-6de2949f4e25-host\") pod \"crc-debug-56h9p\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.548447 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxhtq\" (UniqueName: \"kubernetes.io/projected/d1cc1035-1281-4e76-bb06-6de2949f4e25-kube-api-access-gxhtq\") pod \"crc-debug-56h9p\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:47 crc kubenswrapper[4715]: I1204 14:59:47.655453 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:48 crc kubenswrapper[4715]: I1204 14:59:48.028739 4715 generic.go:334] "Generic (PLEG): container finished" podID="d1cc1035-1281-4e76-bb06-6de2949f4e25" containerID="4eaaeead32434d7f0be07863e0f30804d1e856753c415f1c46b823c02275e288" exitCode=0 Dec 04 14:59:48 crc kubenswrapper[4715]: I1204 14:59:48.028796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-56h9p" event={"ID":"d1cc1035-1281-4e76-bb06-6de2949f4e25","Type":"ContainerDied","Data":"4eaaeead32434d7f0be07863e0f30804d1e856753c415f1c46b823c02275e288"} Dec 04 14:59:48 crc kubenswrapper[4715]: I1204 14:59:48.028832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-56h9p" event={"ID":"d1cc1035-1281-4e76-bb06-6de2949f4e25","Type":"ContainerStarted","Data":"29151bfcc2eef56386e3e7c6827823613e2537418eb238c853ab3fa2853d11c7"} Dec 04 14:59:48 crc kubenswrapper[4715]: I1204 14:59:48.496059 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-56h9p"] Dec 04 14:59:48 crc kubenswrapper[4715]: I1204 14:59:48.504553 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-56h9p"] Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.143887 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.183752 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxhtq\" (UniqueName: \"kubernetes.io/projected/d1cc1035-1281-4e76-bb06-6de2949f4e25-kube-api-access-gxhtq\") pod \"d1cc1035-1281-4e76-bb06-6de2949f4e25\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.184027 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1cc1035-1281-4e76-bb06-6de2949f4e25-host\") pod \"d1cc1035-1281-4e76-bb06-6de2949f4e25\" (UID: \"d1cc1035-1281-4e76-bb06-6de2949f4e25\") " Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.184203 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1cc1035-1281-4e76-bb06-6de2949f4e25-host" (OuterVolumeSpecName: "host") pod "d1cc1035-1281-4e76-bb06-6de2949f4e25" (UID: "d1cc1035-1281-4e76-bb06-6de2949f4e25"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.184683 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1cc1035-1281-4e76-bb06-6de2949f4e25-host\") on node \"crc\" DevicePath \"\"" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.193721 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cc1035-1281-4e76-bb06-6de2949f4e25-kube-api-access-gxhtq" (OuterVolumeSpecName: "kube-api-access-gxhtq") pod "d1cc1035-1281-4e76-bb06-6de2949f4e25" (UID: "d1cc1035-1281-4e76-bb06-6de2949f4e25"). InnerVolumeSpecName "kube-api-access-gxhtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.286890 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxhtq\" (UniqueName: \"kubernetes.io/projected/d1cc1035-1281-4e76-bb06-6de2949f4e25-kube-api-access-gxhtq\") on node \"crc\" DevicePath \"\"" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.651467 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-wx7nq"] Dec 04 14:59:49 crc kubenswrapper[4715]: E1204 14:59:49.652137 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cc1035-1281-4e76-bb06-6de2949f4e25" containerName="container-00" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.652154 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cc1035-1281-4e76-bb06-6de2949f4e25" containerName="container-00" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.652341 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1cc1035-1281-4e76-bb06-6de2949f4e25" containerName="container-00" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.652947 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.694013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-host\") pod \"crc-debug-wx7nq\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.694139 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvl5c\" (UniqueName: \"kubernetes.io/projected/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-kube-api-access-jvl5c\") pod \"crc-debug-wx7nq\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.796443 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-host\") pod \"crc-debug-wx7nq\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.796542 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvl5c\" (UniqueName: \"kubernetes.io/projected/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-kube-api-access-jvl5c\") pod \"crc-debug-wx7nq\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.796998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-host\") pod \"crc-debug-wx7nq\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.817910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvl5c\" (UniqueName: \"kubernetes.io/projected/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-kube-api-access-jvl5c\") pod \"crc-debug-wx7nq\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:49 crc kubenswrapper[4715]: I1204 14:59:49.973519 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:50 crc kubenswrapper[4715]: W1204 14:59:50.020354 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0e80a73_c037_4ea6_83a4_06a15aa3ec25.slice/crio-69d0266562b22a94bcab2f3f2d8e38e630af078de1776dbba8b4f75885b39080 WatchSource:0}: Error finding container 69d0266562b22a94bcab2f3f2d8e38e630af078de1776dbba8b4f75885b39080: Status 404 returned error can't find the container with id 69d0266562b22a94bcab2f3f2d8e38e630af078de1776dbba8b4f75885b39080 Dec 04 14:59:50 crc kubenswrapper[4715]: I1204 14:59:50.046266 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" event={"ID":"a0e80a73-c037-4ea6-83a4-06a15aa3ec25","Type":"ContainerStarted","Data":"69d0266562b22a94bcab2f3f2d8e38e630af078de1776dbba8b4f75885b39080"} Dec 04 14:59:50 crc kubenswrapper[4715]: I1204 14:59:50.047914 4715 scope.go:117] "RemoveContainer" containerID="4eaaeead32434d7f0be07863e0f30804d1e856753c415f1c46b823c02275e288" Dec 04 14:59:50 crc kubenswrapper[4715]: I1204 14:59:50.047937 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-56h9p" Dec 04 14:59:51 crc kubenswrapper[4715]: I1204 14:59:51.057716 4715 generic.go:334] "Generic (PLEG): container finished" podID="a0e80a73-c037-4ea6-83a4-06a15aa3ec25" containerID="5f5d50819c92ed4ea84cf670014b840621450034120e7011b42b7a9977b1ec4c" exitCode=0 Dec 04 14:59:51 crc kubenswrapper[4715]: I1204 14:59:51.057984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" event={"ID":"a0e80a73-c037-4ea6-83a4-06a15aa3ec25","Type":"ContainerDied","Data":"5f5d50819c92ed4ea84cf670014b840621450034120e7011b42b7a9977b1ec4c"} Dec 04 14:59:51 crc kubenswrapper[4715]: I1204 14:59:51.123192 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-wx7nq"] Dec 04 14:59:51 crc kubenswrapper[4715]: I1204 14:59:51.134210 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn5hf/crc-debug-wx7nq"] Dec 04 14:59:51 crc kubenswrapper[4715]: I1204 14:59:51.191238 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cc1035-1281-4e76-bb06-6de2949f4e25" path="/var/lib/kubelet/pods/d1cc1035-1281-4e76-bb06-6de2949f4e25/volumes" Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.186281 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.245436 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-host\") pod \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.245527 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-host" (OuterVolumeSpecName: "host") pod "a0e80a73-c037-4ea6-83a4-06a15aa3ec25" (UID: "a0e80a73-c037-4ea6-83a4-06a15aa3ec25"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.245703 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvl5c\" (UniqueName: \"kubernetes.io/projected/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-kube-api-access-jvl5c\") pod \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\" (UID: \"a0e80a73-c037-4ea6-83a4-06a15aa3ec25\") " Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.246264 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-host\") on node \"crc\" DevicePath \"\"" Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.263155 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-kube-api-access-jvl5c" (OuterVolumeSpecName: "kube-api-access-jvl5c") pod "a0e80a73-c037-4ea6-83a4-06a15aa3ec25" (UID: "a0e80a73-c037-4ea6-83a4-06a15aa3ec25"). InnerVolumeSpecName "kube-api-access-jvl5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:59:52 crc kubenswrapper[4715]: I1204 14:59:52.348245 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvl5c\" (UniqueName: \"kubernetes.io/projected/a0e80a73-c037-4ea6-83a4-06a15aa3ec25-kube-api-access-jvl5c\") on node \"crc\" DevicePath \"\"" Dec 04 14:59:53 crc kubenswrapper[4715]: I1204 14:59:53.077755 4715 scope.go:117] "RemoveContainer" containerID="5f5d50819c92ed4ea84cf670014b840621450034120e7011b42b7a9977b1ec4c" Dec 04 14:59:53 crc kubenswrapper[4715]: I1204 14:59:53.077825 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/crc-debug-wx7nq" Dec 04 14:59:53 crc kubenswrapper[4715]: I1204 14:59:53.192647 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0e80a73-c037-4ea6-83a4-06a15aa3ec25" path="/var/lib/kubelet/pods/a0e80a73-c037-4ea6-83a4-06a15aa3ec25/volumes" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.169834 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp"] Dec 04 15:00:00 crc kubenswrapper[4715]: E1204 15:00:00.170945 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e80a73-c037-4ea6-83a4-06a15aa3ec25" containerName="container-00" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.170963 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e80a73-c037-4ea6-83a4-06a15aa3ec25" containerName="container-00" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.171238 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e80a73-c037-4ea6-83a4-06a15aa3ec25" containerName="container-00" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.188517 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.190734 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.191067 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.210581 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp"] Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.225345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnrfx\" (UniqueName: \"kubernetes.io/projected/9c664cf3-c47f-4e51-a4af-893fb0328423-kube-api-access-rnrfx\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.225510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c664cf3-c47f-4e51-a4af-893fb0328423-config-volume\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.225726 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c664cf3-c47f-4e51-a4af-893fb0328423-secret-volume\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.327377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnrfx\" (UniqueName: \"kubernetes.io/projected/9c664cf3-c47f-4e51-a4af-893fb0328423-kube-api-access-rnrfx\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.327460 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c664cf3-c47f-4e51-a4af-893fb0328423-config-volume\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.327525 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c664cf3-c47f-4e51-a4af-893fb0328423-secret-volume\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.328709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c664cf3-c47f-4e51-a4af-893fb0328423-config-volume\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.343607 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c664cf3-c47f-4e51-a4af-893fb0328423-secret-volume\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.343985 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnrfx\" (UniqueName: \"kubernetes.io/projected/9c664cf3-c47f-4e51-a4af-893fb0328423-kube-api-access-rnrfx\") pod \"collect-profiles-29414340-2vjlp\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:00 crc kubenswrapper[4715]: I1204 15:00:00.550440 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:01 crc kubenswrapper[4715]: I1204 15:00:01.052009 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp"] Dec 04 15:00:01 crc kubenswrapper[4715]: W1204 15:00:01.055081 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c664cf3_c47f_4e51_a4af_893fb0328423.slice/crio-603f9e60fe5b1a88e273b0ecf4111235e7df43fe921265c8d33546c74c6b5263 WatchSource:0}: Error finding container 603f9e60fe5b1a88e273b0ecf4111235e7df43fe921265c8d33546c74c6b5263: Status 404 returned error can't find the container with id 603f9e60fe5b1a88e273b0ecf4111235e7df43fe921265c8d33546c74c6b5263 Dec 04 15:00:01 crc kubenswrapper[4715]: I1204 15:00:01.196360 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" event={"ID":"9c664cf3-c47f-4e51-a4af-893fb0328423","Type":"ContainerStarted","Data":"603f9e60fe5b1a88e273b0ecf4111235e7df43fe921265c8d33546c74c6b5263"} Dec 04 15:00:02 crc kubenswrapper[4715]: I1204 15:00:02.194589 4715 generic.go:334] "Generic (PLEG): container finished" podID="9c664cf3-c47f-4e51-a4af-893fb0328423" containerID="045a2cc6f0d0b052ecf51210ce41d41e863db90053d4cb1ee711536267011a2c" exitCode=0 Dec 04 15:00:02 crc kubenswrapper[4715]: I1204 15:00:02.194645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" event={"ID":"9c664cf3-c47f-4e51-a4af-893fb0328423","Type":"ContainerDied","Data":"045a2cc6f0d0b052ecf51210ce41d41e863db90053d4cb1ee711536267011a2c"} Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.683845 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.793745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnrfx\" (UniqueName: \"kubernetes.io/projected/9c664cf3-c47f-4e51-a4af-893fb0328423-kube-api-access-rnrfx\") pod \"9c664cf3-c47f-4e51-a4af-893fb0328423\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.793842 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c664cf3-c47f-4e51-a4af-893fb0328423-secret-volume\") pod \"9c664cf3-c47f-4e51-a4af-893fb0328423\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.793935 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c664cf3-c47f-4e51-a4af-893fb0328423-config-volume\") pod \"9c664cf3-c47f-4e51-a4af-893fb0328423\" (UID: \"9c664cf3-c47f-4e51-a4af-893fb0328423\") " Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.795098 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c664cf3-c47f-4e51-a4af-893fb0328423-config-volume" (OuterVolumeSpecName: "config-volume") pod "9c664cf3-c47f-4e51-a4af-893fb0328423" (UID: "9c664cf3-c47f-4e51-a4af-893fb0328423"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.799614 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c664cf3-c47f-4e51-a4af-893fb0328423-kube-api-access-rnrfx" (OuterVolumeSpecName: "kube-api-access-rnrfx") pod "9c664cf3-c47f-4e51-a4af-893fb0328423" (UID: "9c664cf3-c47f-4e51-a4af-893fb0328423"). InnerVolumeSpecName "kube-api-access-rnrfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.799856 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c664cf3-c47f-4e51-a4af-893fb0328423-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9c664cf3-c47f-4e51-a4af-893fb0328423" (UID: "9c664cf3-c47f-4e51-a4af-893fb0328423"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.896214 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnrfx\" (UniqueName: \"kubernetes.io/projected/9c664cf3-c47f-4e51-a4af-893fb0328423-kube-api-access-rnrfx\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.896276 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c664cf3-c47f-4e51-a4af-893fb0328423-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:03 crc kubenswrapper[4715]: I1204 15:00:03.896286 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c664cf3-c47f-4e51-a4af-893fb0328423-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:04 crc kubenswrapper[4715]: I1204 15:00:04.214848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" event={"ID":"9c664cf3-c47f-4e51-a4af-893fb0328423","Type":"ContainerDied","Data":"603f9e60fe5b1a88e273b0ecf4111235e7df43fe921265c8d33546c74c6b5263"} Dec 04 15:00:04 crc kubenswrapper[4715]: I1204 15:00:04.214893 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="603f9e60fe5b1a88e273b0ecf4111235e7df43fe921265c8d33546c74c6b5263" Dec 04 15:00:04 crc kubenswrapper[4715]: I1204 15:00:04.214919 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-2vjlp" Dec 04 15:00:04 crc kubenswrapper[4715]: I1204 15:00:04.780115 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr"] Dec 04 15:00:04 crc kubenswrapper[4715]: I1204 15:00:04.792281 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-lw8cr"] Dec 04 15:00:05 crc kubenswrapper[4715]: I1204 15:00:05.198404 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46bcad15-7a3f-4bc0-8ae7-344881de90e5" path="/var/lib/kubelet/pods/46bcad15-7a3f-4bc0-8ae7-344881de90e5/volumes" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.031783 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65b4dd5988-fzzwk_795cce68-d6db-4923-ab86-c474d5085318/barbican-api/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.180115 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65b4dd5988-fzzwk_795cce68-d6db-4923-ab86-c474d5085318/barbican-api-log/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.323840 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f7449dddd-lkzvw_589e9c2f-67f8-4208-aaa3-b391d59b7d01/barbican-keystone-listener/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.376738 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f7449dddd-lkzvw_589e9c2f-67f8-4208-aaa3-b391d59b7d01/barbican-keystone-listener-log/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.527687 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d967d55f-xm4d6_5a34d667-48aa-4bb9-82fa-bb528e9bcc20/barbican-worker/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.551638 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d967d55f-xm4d6_5a34d667-48aa-4bb9-82fa-bb528e9bcc20/barbican-worker-log/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.704555 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt_b6a57886-292b-4dac-bfb6-a51971fcde74/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.830631 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/ceilometer-central-agent/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.914428 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/proxy-httpd/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.918610 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/ceilometer-notification-agent/0.log" Dec 04 15:00:07 crc kubenswrapper[4715]: I1204 15:00:07.969161 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/sg-core/0.log" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.124972 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_87853446-d88c-4d46-9fea-ff90af5617ba/cinder-api-log/0.log" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.162177 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6mjlz"] Dec 04 15:00:08 crc kubenswrapper[4715]: E1204 15:00:08.162572 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c664cf3-c47f-4e51-a4af-893fb0328423" containerName="collect-profiles" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.162583 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c664cf3-c47f-4e51-a4af-893fb0328423" containerName="collect-profiles" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.163649 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c664cf3-c47f-4e51-a4af-893fb0328423" containerName="collect-profiles" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.165046 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.187720 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mjlz"] Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.195796 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6514d4c8-7452-4255-9e08-276daa522ad0-utilities\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.195829 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6514d4c8-7452-4255-9e08-276daa522ad0-catalog-content\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.195875 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhlhg\" (UniqueName: \"kubernetes.io/projected/6514d4c8-7452-4255-9e08-276daa522ad0-kube-api-access-dhlhg\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.237778 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_87853446-d88c-4d46-9fea-ff90af5617ba/cinder-api/0.log" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.299011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6514d4c8-7452-4255-9e08-276daa522ad0-utilities\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.299075 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6514d4c8-7452-4255-9e08-276daa522ad0-catalog-content\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.299146 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhlhg\" (UniqueName: \"kubernetes.io/projected/6514d4c8-7452-4255-9e08-276daa522ad0-kube-api-access-dhlhg\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.299613 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6514d4c8-7452-4255-9e08-276daa522ad0-utilities\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.299672 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6514d4c8-7452-4255-9e08-276daa522ad0-catalog-content\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.322889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhlhg\" (UniqueName: \"kubernetes.io/projected/6514d4c8-7452-4255-9e08-276daa522ad0-kube-api-access-dhlhg\") pod \"certified-operators-6mjlz\" (UID: \"6514d4c8-7452-4255-9e08-276daa522ad0\") " pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.461328 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6816724d-3e76-4832-9e96-ab6ef342c354/probe/0.log" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.502337 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6816724d-3e76-4832-9e96-ab6ef342c354/cinder-scheduler/0.log" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.521319 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.621277 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6t47w_7f4cab97-fa52-47c4-ad8e-7469556526cc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:08 crc kubenswrapper[4715]: I1204 15:00:08.889706 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq_337aa155-3f3f-4e84-a350-29d27ef13260/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.103061 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mjlz"] Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.191078 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-rbrt6_dfe175eb-a594-4a23-a057-b3cbc53cad3e/init/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.260629 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mjlz" event={"ID":"6514d4c8-7452-4255-9e08-276daa522ad0","Type":"ContainerStarted","Data":"cd0f66f5ba09f39dc9f5767a6e05256e670a6287fc60b93c10c238d02c869378"} Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.345637 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-rbrt6_dfe175eb-a594-4a23-a057-b3cbc53cad3e/init/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.361578 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-rbrt6_dfe175eb-a594-4a23-a057-b3cbc53cad3e/dnsmasq-dns/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.432686 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-8dv88_a8a8da4d-f578-4428-8b74-9d4efd5f78ee/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.581442 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92c1ce5a-911d-4567-8d23-11777a6df443/glance-httpd/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.616759 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92c1ce5a-911d-4567-8d23-11777a6df443/glance-log/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.789141 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3bea4af7-30c3-4278-99d7-01f1c604a89c/glance-httpd/0.log" Dec 04 15:00:09 crc kubenswrapper[4715]: I1204 15:00:09.808064 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3bea4af7-30c3-4278-99d7-01f1c604a89c/glance-log/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.007005 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5b798c69c6-g6fvm_8552f780-12e2-4eeb-bbc5-9ef637e9278f/horizon/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.118676 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv_d394ffb7-2675-4316-8388-cee93c9a37bc/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.256344 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5b798c69c6-g6fvm_8552f780-12e2-4eeb-bbc5-9ef637e9278f/horizon-log/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.294428 4715 generic.go:334] "Generic (PLEG): container finished" podID="6514d4c8-7452-4255-9e08-276daa522ad0" containerID="e0a505ca9b3a389a6511215b4170b5f26b6c95d46182ac233f19a74a786d1759" exitCode=0 Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.294480 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mjlz" event={"ID":"6514d4c8-7452-4255-9e08-276daa522ad0","Type":"ContainerDied","Data":"e0a505ca9b3a389a6511215b4170b5f26b6c95d46182ac233f19a74a786d1759"} Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.297086 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.308813 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6l4js_b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.504736 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b0d685d3-1f35-497c-986b-14d7407716d7/kube-state-metrics/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.553323 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8wb78"] Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.556394 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.581486 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wb78"] Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.647891 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx5b5\" (UniqueName: \"kubernetes.io/projected/7fdee0cf-4922-4c43-8b17-c2de5a379272-kube-api-access-sx5b5\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.648419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-catalog-content\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.648659 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-utilities\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.649983 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b79f64b59-wfklg_a412f2fc-2452-41bc-9670-d0db2350f145/keystone-api/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.750899 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-utilities\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.751291 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx5b5\" (UniqueName: \"kubernetes.io/projected/7fdee0cf-4922-4c43-8b17-c2de5a379272-kube-api-access-sx5b5\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.751364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-catalog-content\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.751385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-utilities\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.751693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-catalog-content\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.778418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx5b5\" (UniqueName: \"kubernetes.io/projected/7fdee0cf-4922-4c43-8b17-c2de5a379272-kube-api-access-sx5b5\") pod \"redhat-marketplace-8wb78\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.864793 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt_4a13c152-6be0-4261-ae2f-3b7b3458005a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:10 crc kubenswrapper[4715]: I1204 15:00:10.913781 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:11 crc kubenswrapper[4715]: I1204 15:00:11.369294 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58655b5d9c-m5lkw_4e9c813e-1ad1-40fd-b015-a312efff6d9e/neutron-api/0.log" Dec 04 15:00:11 crc kubenswrapper[4715]: I1204 15:00:11.432269 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58655b5d9c-m5lkw_4e9c813e-1ad1-40fd-b015-a312efff6d9e/neutron-httpd/0.log" Dec 04 15:00:11 crc kubenswrapper[4715]: I1204 15:00:11.460335 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wb78"] Dec 04 15:00:11 crc kubenswrapper[4715]: I1204 15:00:11.460808 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl_e8a7fc4f-4f90-4023-98af-a9613dbce1cb/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.101473 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9ec54636-c541-4292-b5b9-7a6215b8828e/nova-api-log/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.121640 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_02266663-d1a7-4968-ab1e-e4bb6b4a9b21/nova-cell0-conductor-conductor/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.314758 4715 generic.go:334] "Generic (PLEG): container finished" podID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerID="3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc" exitCode=0 Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.314804 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wb78" event={"ID":"7fdee0cf-4922-4c43-8b17-c2de5a379272","Type":"ContainerDied","Data":"3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc"} Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.314835 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wb78" event={"ID":"7fdee0cf-4922-4c43-8b17-c2de5a379272","Type":"ContainerStarted","Data":"869ef40fe0b94b59af15009b2f13a933dc32736067ab53843504970a41fbdaea"} Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.342908 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0246a68d-8dff-43c6-9efa-fef70ae0919b/nova-cell1-conductor-conductor/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.396580 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9ec54636-c541-4292-b5b9-7a6215b8828e/nova-api-api/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.484875 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9701ac27-1f36-4d70-a5ac-2197ea8880f6/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.736677 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-gjztt_df89956f-dd09-4cd1-bc58-d57f555d2870/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:12 crc kubenswrapper[4715]: I1204 15:00:12.866182 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_62b5d096-e3b9-4fae-a697-d5a4b693514d/nova-metadata-log/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.128879 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e478a0b5-cea1-4639-a645-9660524ff9bd/mysql-bootstrap/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.195653 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce/nova-scheduler-scheduler/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.358409 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e478a0b5-cea1-4639-a645-9660524ff9bd/galera/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.370826 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e478a0b5-cea1-4639-a645-9660524ff9bd/mysql-bootstrap/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.590701 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7d666280-e175-4f31-8e11-7af2d5e27e5a/mysql-bootstrap/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.767501 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7d666280-e175-4f31-8e11-7af2d5e27e5a/mysql-bootstrap/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.832657 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7d666280-e175-4f31-8e11-7af2d5e27e5a/galera/0.log" Dec 04 15:00:13 crc kubenswrapper[4715]: I1204 15:00:13.952197 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72/openstackclient/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.034087 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6k4jg_00743321-f084-4e95-b47d-1b21332be3d5/ovn-controller/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.217324 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-tw9p4_44a78956-6eb6-43e2-98f1-2e2c15342fd8/openstack-network-exporter/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.341006 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_62b5d096-e3b9-4fae-a697-d5a4b693514d/nova-metadata-metadata/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.345303 4715 generic.go:334] "Generic (PLEG): container finished" podID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerID="13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f" exitCode=0 Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.345358 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wb78" event={"ID":"7fdee0cf-4922-4c43-8b17-c2de5a379272","Type":"ContainerDied","Data":"13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f"} Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.429272 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovsdb-server-init/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.623644 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovsdb-server-init/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.705051 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovs-vswitchd/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.713249 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovsdb-server/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.890373 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6qf62_d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:14 crc kubenswrapper[4715]: I1204 15:00:14.901427 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_57e4b94e-20df-457c-81e8-9c114b18bbe6/openstack-network-exporter/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.046926 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_57e4b94e-20df-457c-81e8-9c114b18bbe6/ovn-northd/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.106695 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d693b193-c70b-4654-828d-5cd4744bee62/openstack-network-exporter/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.178815 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d693b193-c70b-4654-828d-5cd4744bee62/ovsdbserver-nb/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.346650 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f01e0a29-0003-41df-a5c9-0ede7a249acc/openstack-network-exporter/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.347493 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f01e0a29-0003-41df-a5c9-0ede7a249acc/ovsdbserver-sb/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.643371 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65f6dc6d96-gj9d4_aad8392c-fc50-4cd5-af7c-9d186021d336/placement-api/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.715457 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65f6dc6d96-gj9d4_aad8392c-fc50-4cd5-af7c-9d186021d336/placement-log/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.736526 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_45d230ae-3ae5-4d87-8979-bbd1fd8651cd/setup-container/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.947511 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_45d230ae-3ae5-4d87-8979-bbd1fd8651cd/rabbitmq/0.log" Dec 04 15:00:15 crc kubenswrapper[4715]: I1204 15:00:15.992999 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_45d230ae-3ae5-4d87-8979-bbd1fd8651cd/setup-container/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.039258 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_72cfc89d-7970-45e9-be5f-3f5f5d3f520d/setup-container/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.270903 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_72cfc89d-7970-45e9-be5f-3f5f5d3f520d/setup-container/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.423134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj_baed77e4-c68b-4663-b4c2-0f0de712d963/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.480218 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_72cfc89d-7970-45e9-be5f-3f5f5d3f520d/rabbitmq/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.657395 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-plght_9542e2ff-8723-4d1d-b17c-277976018a11/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.700406 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt_76dfdae9-34be-48e3-a220-4322d19e7260/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.917837 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-fpzs7_d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:16 crc kubenswrapper[4715]: I1204 15:00:16.919054 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-4wlm4_5058ed17-7d86-4b90-a41b-668b6407054a/ssh-known-hosts-edpm-deployment/0.log" Dec 04 15:00:17 crc kubenswrapper[4715]: I1204 15:00:17.251248 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-64bbd97b75-ssj5n_861e33f9-3559-444f-8120-f439948fc774/proxy-server/0.log" Dec 04 15:00:17 crc kubenswrapper[4715]: I1204 15:00:17.550608 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-64bbd97b75-ssj5n_861e33f9-3559-444f-8120-f439948fc774/proxy-httpd/0.log" Dec 04 15:00:17 crc kubenswrapper[4715]: I1204 15:00:17.878646 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-auditor/0.log" Dec 04 15:00:17 crc kubenswrapper[4715]: I1204 15:00:17.931616 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5r72w_2386cbfc-2eba-4ace-a56a-0e203b7b9229/swift-ring-rebalance/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.121285 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-reaper/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.181425 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-auditor/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.214408 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-replicator/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.461466 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-server/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.670178 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-auditor/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.779298 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-updater/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.794127 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-server/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.818168 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-replicator/0.log" Dec 04 15:00:18 crc kubenswrapper[4715]: I1204 15:00:18.840060 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-expirer/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.009690 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-server/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.025276 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-updater/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.087285 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-replicator/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.133550 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/rsync/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.230690 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/swift-recon-cron/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.350941 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz_2fa80717-8afa-4d1a-ac4f-4b807445f638/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:19 crc kubenswrapper[4715]: I1204 15:00:19.923468 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1314b612-6502-4cef-ade7-f05971b781a2/tempest-tests-tempest-tests-runner/0.log" Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.285835 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_212e2726-336d-42d9-942a-36642b0c1c98/test-operator-logs-container/0.log" Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.397791 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cq24p_ee373f57-aa34-4c0a-86d4-5113c10984df/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.412664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wb78" event={"ID":"7fdee0cf-4922-4c43-8b17-c2de5a379272","Type":"ContainerStarted","Data":"4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a"} Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.415763 4715 generic.go:334] "Generic (PLEG): container finished" podID="6514d4c8-7452-4255-9e08-276daa522ad0" containerID="f1d5289e9b33ea0ab134d599582c66e9ea2437679801737364e7f1696e2e9554" exitCode=0 Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.415807 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mjlz" event={"ID":"6514d4c8-7452-4255-9e08-276daa522ad0","Type":"ContainerDied","Data":"f1d5289e9b33ea0ab134d599582c66e9ea2437679801737364e7f1696e2e9554"} Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.442470 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8wb78" podStartSLOduration=3.5869983 podStartE2EDuration="10.442448502s" podCreationTimestamp="2025-12-04 15:00:10 +0000 UTC" firstStartedPulling="2025-12-04 15:00:12.317732155 +0000 UTC m=+3769.386450370" lastFinishedPulling="2025-12-04 15:00:19.173182357 +0000 UTC m=+3776.241900572" observedRunningTime="2025-12-04 15:00:20.435483241 +0000 UTC m=+3777.504201486" watchObservedRunningTime="2025-12-04 15:00:20.442448502 +0000 UTC m=+3777.511166717" Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.913931 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:20 crc kubenswrapper[4715]: I1204 15:00:20.914056 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:21 crc kubenswrapper[4715]: I1204 15:00:21.428204 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mjlz" event={"ID":"6514d4c8-7452-4255-9e08-276daa522ad0","Type":"ContainerStarted","Data":"8b654877e63eaf9c8dbdb6aa16d550b6aa76db5cbd5482f557f4071bd1e8e77d"} Dec 04 15:00:21 crc kubenswrapper[4715]: I1204 15:00:21.451174 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6mjlz" podStartSLOduration=2.668586073 podStartE2EDuration="13.45115672s" podCreationTimestamp="2025-12-04 15:00:08 +0000 UTC" firstStartedPulling="2025-12-04 15:00:10.296760849 +0000 UTC m=+3767.365479064" lastFinishedPulling="2025-12-04 15:00:21.079331496 +0000 UTC m=+3778.148049711" observedRunningTime="2025-12-04 15:00:21.446648547 +0000 UTC m=+3778.515366762" watchObservedRunningTime="2025-12-04 15:00:21.45115672 +0000 UTC m=+3778.519874935" Dec 04 15:00:21 crc kubenswrapper[4715]: I1204 15:00:21.978840 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-8wb78" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="registry-server" probeResult="failure" output=< Dec 04 15:00:21 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 04 15:00:21 crc kubenswrapper[4715]: > Dec 04 15:00:28 crc kubenswrapper[4715]: I1204 15:00:28.521496 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:28 crc kubenswrapper[4715]: I1204 15:00:28.522237 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:28 crc kubenswrapper[4715]: I1204 15:00:28.576466 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:28 crc kubenswrapper[4715]: I1204 15:00:28.667364 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_91e88f07-d0d0-46d4-9e6a-a973b37b8a72/memcached/0.log" Dec 04 15:00:29 crc kubenswrapper[4715]: I1204 15:00:29.557704 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6mjlz" Dec 04 15:00:29 crc kubenswrapper[4715]: I1204 15:00:29.628498 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mjlz"] Dec 04 15:00:29 crc kubenswrapper[4715]: I1204 15:00:29.673080 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2fvf6"] Dec 04 15:00:29 crc kubenswrapper[4715]: I1204 15:00:29.673430 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2fvf6" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="registry-server" containerID="cri-o://34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18" gracePeriod=2 Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.174593 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.314301 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-catalog-content\") pod \"4968998b-fada-4949-beb1-982de165dc3c\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.314420 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-utilities\") pod \"4968998b-fada-4949-beb1-982de165dc3c\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.314597 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjwrp\" (UniqueName: \"kubernetes.io/projected/4968998b-fada-4949-beb1-982de165dc3c-kube-api-access-xjwrp\") pod \"4968998b-fada-4949-beb1-982de165dc3c\" (UID: \"4968998b-fada-4949-beb1-982de165dc3c\") " Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.317513 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-utilities" (OuterVolumeSpecName: "utilities") pod "4968998b-fada-4949-beb1-982de165dc3c" (UID: "4968998b-fada-4949-beb1-982de165dc3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.322572 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4968998b-fada-4949-beb1-982de165dc3c-kube-api-access-xjwrp" (OuterVolumeSpecName: "kube-api-access-xjwrp") pod "4968998b-fada-4949-beb1-982de165dc3c" (UID: "4968998b-fada-4949-beb1-982de165dc3c"). InnerVolumeSpecName "kube-api-access-xjwrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.377449 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4968998b-fada-4949-beb1-982de165dc3c" (UID: "4968998b-fada-4949-beb1-982de165dc3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.416757 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.416806 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4968998b-fada-4949-beb1-982de165dc3c-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.416821 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjwrp\" (UniqueName: \"kubernetes.io/projected/4968998b-fada-4949-beb1-982de165dc3c-kube-api-access-xjwrp\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.514422 4715 generic.go:334] "Generic (PLEG): container finished" podID="4968998b-fada-4949-beb1-982de165dc3c" containerID="34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18" exitCode=0 Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.514465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2fvf6" event={"ID":"4968998b-fada-4949-beb1-982de165dc3c","Type":"ContainerDied","Data":"34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18"} Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.514524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2fvf6" event={"ID":"4968998b-fada-4949-beb1-982de165dc3c","Type":"ContainerDied","Data":"9ac5ed4c66ecd8e8a00e9a52ce5c688e7eef85de547c12443e25a755f512206f"} Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.514521 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2fvf6" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.514572 4715 scope.go:117] "RemoveContainer" containerID="34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.539466 4715 scope.go:117] "RemoveContainer" containerID="bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.544102 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2fvf6"] Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.552427 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2fvf6"] Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.564301 4715 scope.go:117] "RemoveContainer" containerID="e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.608663 4715 scope.go:117] "RemoveContainer" containerID="34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18" Dec 04 15:00:30 crc kubenswrapper[4715]: E1204 15:00:30.609366 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18\": container with ID starting with 34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18 not found: ID does not exist" containerID="34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.609396 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18"} err="failed to get container status \"34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18\": rpc error: code = NotFound desc = could not find container \"34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18\": container with ID starting with 34cfea31893b053f2169ffdcb9ed25faf9ec144b3a278568954ea55d09432c18 not found: ID does not exist" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.609415 4715 scope.go:117] "RemoveContainer" containerID="bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485" Dec 04 15:00:30 crc kubenswrapper[4715]: E1204 15:00:30.609761 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485\": container with ID starting with bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485 not found: ID does not exist" containerID="bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.609869 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485"} err="failed to get container status \"bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485\": rpc error: code = NotFound desc = could not find container \"bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485\": container with ID starting with bf60c05bf3c59f1b0235bd7bbf6d93469bdf3e239f523cad9afc6c00af39b485 not found: ID does not exist" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.609972 4715 scope.go:117] "RemoveContainer" containerID="e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d" Dec 04 15:00:30 crc kubenswrapper[4715]: E1204 15:00:30.610326 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d\": container with ID starting with e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d not found: ID does not exist" containerID="e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.610360 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d"} err="failed to get container status \"e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d\": rpc error: code = NotFound desc = could not find container \"e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d\": container with ID starting with e4176f3b02aeea34cdc4e1c0528fdb63f752dc12d08ef986ae3ec73bb3f1c29d not found: ID does not exist" Dec 04 15:00:30 crc kubenswrapper[4715]: I1204 15:00:30.972980 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:31 crc kubenswrapper[4715]: I1204 15:00:31.028943 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:31 crc kubenswrapper[4715]: I1204 15:00:31.190205 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4968998b-fada-4949-beb1-982de165dc3c" path="/var/lib/kubelet/pods/4968998b-fada-4949-beb1-982de165dc3c/volumes" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.008892 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wb78"] Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.009376 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8wb78" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="registry-server" containerID="cri-o://4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a" gracePeriod=2 Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.496007 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.556062 4715 generic.go:334] "Generic (PLEG): container finished" podID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerID="4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a" exitCode=0 Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.556102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wb78" event={"ID":"7fdee0cf-4922-4c43-8b17-c2de5a379272","Type":"ContainerDied","Data":"4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a"} Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.556130 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wb78" event={"ID":"7fdee0cf-4922-4c43-8b17-c2de5a379272","Type":"ContainerDied","Data":"869ef40fe0b94b59af15009b2f13a933dc32736067ab53843504970a41fbdaea"} Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.556148 4715 scope.go:117] "RemoveContainer" containerID="4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.556283 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wb78" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.573866 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx5b5\" (UniqueName: \"kubernetes.io/projected/7fdee0cf-4922-4c43-8b17-c2de5a379272-kube-api-access-sx5b5\") pod \"7fdee0cf-4922-4c43-8b17-c2de5a379272\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.573933 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-catalog-content\") pod \"7fdee0cf-4922-4c43-8b17-c2de5a379272\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.574003 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-utilities\") pod \"7fdee0cf-4922-4c43-8b17-c2de5a379272\" (UID: \"7fdee0cf-4922-4c43-8b17-c2de5a379272\") " Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.574949 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-utilities" (OuterVolumeSpecName: "utilities") pod "7fdee0cf-4922-4c43-8b17-c2de5a379272" (UID: "7fdee0cf-4922-4c43-8b17-c2de5a379272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.579272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fdee0cf-4922-4c43-8b17-c2de5a379272-kube-api-access-sx5b5" (OuterVolumeSpecName: "kube-api-access-sx5b5") pod "7fdee0cf-4922-4c43-8b17-c2de5a379272" (UID: "7fdee0cf-4922-4c43-8b17-c2de5a379272"). InnerVolumeSpecName "kube-api-access-sx5b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.584788 4715 scope.go:117] "RemoveContainer" containerID="13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.591889 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fdee0cf-4922-4c43-8b17-c2de5a379272" (UID: "7fdee0cf-4922-4c43-8b17-c2de5a379272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.647238 4715 scope.go:117] "RemoveContainer" containerID="3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.675616 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx5b5\" (UniqueName: \"kubernetes.io/projected/7fdee0cf-4922-4c43-8b17-c2de5a379272-kube-api-access-sx5b5\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.675648 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.675657 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fdee0cf-4922-4c43-8b17-c2de5a379272-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.688962 4715 scope.go:117] "RemoveContainer" containerID="4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a" Dec 04 15:00:33 crc kubenswrapper[4715]: E1204 15:00:33.689447 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a\": container with ID starting with 4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a not found: ID does not exist" containerID="4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.689512 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a"} err="failed to get container status \"4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a\": rpc error: code = NotFound desc = could not find container \"4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a\": container with ID starting with 4e75ca362cace97fe95557f9cff5dc4252b614d06f008a41f8f6f7ffc0181d1a not found: ID does not exist" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.689547 4715 scope.go:117] "RemoveContainer" containerID="13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f" Dec 04 15:00:33 crc kubenswrapper[4715]: E1204 15:00:33.689912 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f\": container with ID starting with 13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f not found: ID does not exist" containerID="13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.689969 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f"} err="failed to get container status \"13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f\": rpc error: code = NotFound desc = could not find container \"13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f\": container with ID starting with 13b4164b3939af6ffd4f2adf7a2fbde3ffc5e8635e220e0d13bda4de922d8b9f not found: ID does not exist" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.690012 4715 scope.go:117] "RemoveContainer" containerID="3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc" Dec 04 15:00:33 crc kubenswrapper[4715]: E1204 15:00:33.691711 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc\": container with ID starting with 3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc not found: ID does not exist" containerID="3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.691767 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc"} err="failed to get container status \"3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc\": rpc error: code = NotFound desc = could not find container \"3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc\": container with ID starting with 3ef3743c14dd853576b27c419bfd34a26d5652d447261b3c6218e7a8e35d62cc not found: ID does not exist" Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.894411 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wb78"] Dec 04 15:00:33 crc kubenswrapper[4715]: I1204 15:00:33.910338 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wb78"] Dec 04 15:00:35 crc kubenswrapper[4715]: I1204 15:00:35.198766 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" path="/var/lib/kubelet/pods/7fdee0cf-4922-4c43-8b17-c2de5a379272/volumes" Dec 04 15:00:38 crc kubenswrapper[4715]: I1204 15:00:38.759643 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:00:38 crc kubenswrapper[4715]: I1204 15:00:38.761106 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:00:45 crc kubenswrapper[4715]: I1204 15:00:45.344773 4715 scope.go:117] "RemoveContainer" containerID="3cf61e5e6767afdc90f9ce2d205b45315fe6ad2449aee48f6790d497229b2482" Dec 04 15:00:46 crc kubenswrapper[4715]: I1204 15:00:46.955366 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/util/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.513507 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/util/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.530740 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/pull/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.582774 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/pull/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.796478 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/pull/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.800374 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/util/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.807868 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/extract/0.log" Dec 04 15:00:47 crc kubenswrapper[4715]: I1204 15:00:47.978274 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-snfrf_d7fc7c62-f1ba-4fcc-9b82-b980988c90cf/kube-rbac-proxy/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.085133 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fdgb9_6efc64e6-15fa-4c60-9a4e-18d326e4c759/kube-rbac-proxy/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.100442 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-snfrf_d7fc7c62-f1ba-4fcc-9b82-b980988c90cf/manager/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.244665 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fdgb9_6efc64e6-15fa-4c60-9a4e-18d326e4c759/manager/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.302365 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-5m69r_4b1a2407-c585-4f22-95db-a2331a69623d/kube-rbac-proxy/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.327655 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-5m69r_4b1a2407-c585-4f22-95db-a2331a69623d/manager/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.485651 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-grcg8_96aa918a-abfb-44f6-bfa3-9849ef0f51e4/kube-rbac-proxy/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.580826 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-grcg8_96aa918a-abfb-44f6-bfa3-9849ef0f51e4/manager/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.686384 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gv5kg_fbc62f1d-2aa8-4b6b-8693-a16c349461d3/kube-rbac-proxy/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.725796 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gv5kg_fbc62f1d-2aa8-4b6b-8693-a16c349461d3/manager/0.log" Dec 04 15:00:48 crc kubenswrapper[4715]: I1204 15:00:48.772603 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5glkm_81510632-c76c-4465-8dec-d8d9c3e8c88d/kube-rbac-proxy/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.015518 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5glkm_81510632-c76c-4465-8dec-d8d9c3e8c88d/manager/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.061727 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-f8cgr_8a9a0373-b39e-496a-aba7-456bddcf3a65/kube-rbac-proxy/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.279075 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-cw997_ec0c61ca-d902-4703-85bc-87b55736fe78/kube-rbac-proxy/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.285638 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-f8cgr_8a9a0373-b39e-496a-aba7-456bddcf3a65/manager/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.311432 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-cw997_ec0c61ca-d902-4703-85bc-87b55736fe78/manager/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.466722 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-rtj2r_9c67b2dc-8ac4-4dd0-af9f-19e55d368092/kube-rbac-proxy/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.588273 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-rtj2r_9c67b2dc-8ac4-4dd0-af9f-19e55d368092/manager/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.703238 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-hhjwx_45a4ad9a-e82a-466a-954b-12127626b577/kube-rbac-proxy/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.708977 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-hhjwx_45a4ad9a-e82a-466a-954b-12127626b577/manager/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.787402 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-xncpr_5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd/kube-rbac-proxy/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.927007 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-xncpr_5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd/manager/0.log" Dec 04 15:00:49 crc kubenswrapper[4715]: I1204 15:00:49.977103 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6hbp2_3521ae52-a647-4190-8bbd-1100bb31dae5/kube-rbac-proxy/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.036608 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6hbp2_3521ae52-a647-4190-8bbd-1100bb31dae5/manager/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.228679 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vnndg_ab0e965e-1d56-4b31-8cf1-6a21892d38ca/kube-rbac-proxy/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.303724 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vnndg_ab0e965e-1d56-4b31-8cf1-6a21892d38ca/manager/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.526143 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dg7m6_4238708b-a19f-485f-a8a7-66b58aac168f/kube-rbac-proxy/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.576928 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dg7m6_4238708b-a19f-485f-a8a7-66b58aac168f/manager/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.657844 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w_73189a74-45fd-4ea8-92ce-bfa038531159/kube-rbac-proxy/0.log" Dec 04 15:00:50 crc kubenswrapper[4715]: I1204 15:00:50.725143 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w_73189a74-45fd-4ea8-92ce-bfa038531159/manager/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.153901 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pnqdc_f7e73b69-ec29-4bdb-9374-25916f8df390/registry-server/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.172740 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6cf7bc7f5d-kzpwj_21b74fe1-833a-4744-b868-6b09c3e93458/operator/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.304755 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-wwc57_84876a00-902b-41f4-af69-f39d588e34fe/kube-rbac-proxy/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.415648 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-w2drg_9d367397-9c5c-4ef7-bf95-91d0a453cf04/kube-rbac-proxy/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.537375 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-wwc57_84876a00-902b-41f4-af69-f39d588e34fe/manager/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.654707 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-w2drg_9d367397-9c5c-4ef7-bf95-91d0a453cf04/manager/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.733324 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9g592_43170033-ddda-4fdf-b0e7-8d3b6d15e4f9/operator/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.949738 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qq7m7_6d5d27a6-eba8-4699-a415-e0e8c283d9cf/kube-rbac-proxy/0.log" Dec 04 15:00:51 crc kubenswrapper[4715]: I1204 15:00:51.955309 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qq7m7_6d5d27a6-eba8-4699-a415-e0e8c283d9cf/manager/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.080545 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-xc2z5_cec0926c-ac5c-4ef2-aefc-a81c7885d810/kube-rbac-proxy/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.125646 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b8c75fd85-txxk5_00ef5230-a1c9-4609-ba71-e90760f29a15/manager/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.221857 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-xc2z5_cec0926c-ac5c-4ef2-aefc-a81c7885d810/manager/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.262148 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-sg77c_88013762-6dd2-4994-b34e-21c89ed5afee/kube-rbac-proxy/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.322852 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-sg77c_88013762-6dd2-4994-b34e-21c89ed5afee/manager/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.423737 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-8ztmn_d58523b1-200c-4b54-ab7c-fab0ee9ff571/kube-rbac-proxy/0.log" Dec 04 15:00:52 crc kubenswrapper[4715]: I1204 15:00:52.464847 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-8ztmn_d58523b1-200c-4b54-ab7c-fab0ee9ff571/manager/0.log" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.151595 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29414341-fhbsf"] Dec 04 15:01:00 crc kubenswrapper[4715]: E1204 15:01:00.153333 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="extract-utilities" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153349 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="extract-utilities" Dec 04 15:01:00 crc kubenswrapper[4715]: E1204 15:01:00.153358 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="registry-server" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153366 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="registry-server" Dec 04 15:01:00 crc kubenswrapper[4715]: E1204 15:01:00.153377 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="extract-content" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153384 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="extract-content" Dec 04 15:01:00 crc kubenswrapper[4715]: E1204 15:01:00.153394 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="extract-utilities" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153399 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="extract-utilities" Dec 04 15:01:00 crc kubenswrapper[4715]: E1204 15:01:00.153433 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="registry-server" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153439 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="registry-server" Dec 04 15:01:00 crc kubenswrapper[4715]: E1204 15:01:00.153459 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="extract-content" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153464 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="extract-content" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153647 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4968998b-fada-4949-beb1-982de165dc3c" containerName="registry-server" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.153676 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdee0cf-4922-4c43-8b17-c2de5a379272" containerName="registry-server" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.154335 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.167749 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414341-fhbsf"] Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.242220 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-combined-ca-bundle\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.242389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxzqr\" (UniqueName: \"kubernetes.io/projected/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-kube-api-access-zxzqr\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.242442 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-fernet-keys\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.242550 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-config-data\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.343872 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-fernet-keys\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.344341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-config-data\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.344528 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-combined-ca-bundle\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.344712 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxzqr\" (UniqueName: \"kubernetes.io/projected/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-kube-api-access-zxzqr\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.351342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-combined-ca-bundle\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.351841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-config-data\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.356949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-fernet-keys\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.363953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxzqr\" (UniqueName: \"kubernetes.io/projected/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-kube-api-access-zxzqr\") pod \"keystone-cron-29414341-fhbsf\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:00 crc kubenswrapper[4715]: I1204 15:01:00.487572 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:01 crc kubenswrapper[4715]: I1204 15:01:01.013589 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414341-fhbsf"] Dec 04 15:01:01 crc kubenswrapper[4715]: I1204 15:01:01.799521 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-fhbsf" event={"ID":"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0","Type":"ContainerStarted","Data":"76525ec6ecba82adbdb7dce0915be88aaf0703ccfbe64096541037a14c1f97b9"} Dec 04 15:01:01 crc kubenswrapper[4715]: I1204 15:01:01.799960 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-fhbsf" event={"ID":"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0","Type":"ContainerStarted","Data":"83f00814e7c9c6ea27fc7dda863fa9cd47246f0fe3e3cf3161b51be66622881e"} Dec 04 15:01:01 crc kubenswrapper[4715]: I1204 15:01:01.835328 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29414341-fhbsf" podStartSLOduration=1.835299977 podStartE2EDuration="1.835299977s" podCreationTimestamp="2025-12-04 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:01:01.813577572 +0000 UTC m=+3818.882295787" watchObservedRunningTime="2025-12-04 15:01:01.835299977 +0000 UTC m=+3818.904018192" Dec 04 15:01:03 crc kubenswrapper[4715]: I1204 15:01:03.820013 4715 generic.go:334] "Generic (PLEG): container finished" podID="1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" containerID="76525ec6ecba82adbdb7dce0915be88aaf0703ccfbe64096541037a14c1f97b9" exitCode=0 Dec 04 15:01:03 crc kubenswrapper[4715]: I1204 15:01:03.820110 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-fhbsf" event={"ID":"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0","Type":"ContainerDied","Data":"76525ec6ecba82adbdb7dce0915be88aaf0703ccfbe64096541037a14c1f97b9"} Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.191324 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.385123 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-combined-ca-bundle\") pod \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.385305 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxzqr\" (UniqueName: \"kubernetes.io/projected/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-kube-api-access-zxzqr\") pod \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.385347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-config-data\") pod \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.385403 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-fernet-keys\") pod \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\" (UID: \"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0\") " Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.397222 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" (UID: "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.397278 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-kube-api-access-zxzqr" (OuterVolumeSpecName: "kube-api-access-zxzqr") pod "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" (UID: "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0"). InnerVolumeSpecName "kube-api-access-zxzqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.422098 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" (UID: "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.456698 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-config-data" (OuterVolumeSpecName: "config-data") pod "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" (UID: "1e0a90fb-c84b-45ee-ae9f-7975ff939ef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.487720 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxzqr\" (UniqueName: \"kubernetes.io/projected/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-kube-api-access-zxzqr\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.487759 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.487775 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.487787 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0a90fb-c84b-45ee-ae9f-7975ff939ef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.865347 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-fhbsf" event={"ID":"1e0a90fb-c84b-45ee-ae9f-7975ff939ef0","Type":"ContainerDied","Data":"83f00814e7c9c6ea27fc7dda863fa9cd47246f0fe3e3cf3161b51be66622881e"} Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.865409 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83f00814e7c9c6ea27fc7dda863fa9cd47246f0fe3e3cf3161b51be66622881e" Dec 04 15:01:05 crc kubenswrapper[4715]: I1204 15:01:05.865497 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-fhbsf" Dec 04 15:01:08 crc kubenswrapper[4715]: I1204 15:01:08.758256 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:01:08 crc kubenswrapper[4715]: I1204 15:01:08.758559 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:01:10 crc kubenswrapper[4715]: I1204 15:01:10.958697 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dvrxw_288f2069-b715-4316-ac68-c075546d04c4/control-plane-machine-set-operator/0.log" Dec 04 15:01:11 crc kubenswrapper[4715]: I1204 15:01:11.127742 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c8mr7_0cf11e59-8054-46a5-8a2d-f7c6434a6a91/kube-rbac-proxy/0.log" Dec 04 15:01:11 crc kubenswrapper[4715]: I1204 15:01:11.169161 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c8mr7_0cf11e59-8054-46a5-8a2d-f7c6434a6a91/machine-api-operator/0.log" Dec 04 15:01:23 crc kubenswrapper[4715]: I1204 15:01:23.615152 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-8nphx_70e6d48d-4b22-42ea-9016-a918b6889d78/cert-manager-controller/0.log" Dec 04 15:01:23 crc kubenswrapper[4715]: I1204 15:01:23.791306 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4tnrx_f4374824-e956-48c8-a4e5-52c55c7486e6/cert-manager-cainjector/0.log" Dec 04 15:01:23 crc kubenswrapper[4715]: I1204 15:01:23.822914 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qbkqd_288f92f5-abcf-40e4-937a-93f6f5ba002b/cert-manager-webhook/0.log" Dec 04 15:01:35 crc kubenswrapper[4715]: I1204 15:01:35.392294 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-pg74q_29cbcfcf-ff1d-4605-ac6e-d7c609de5911/nmstate-console-plugin/0.log" Dec 04 15:01:35 crc kubenswrapper[4715]: I1204 15:01:35.512074 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lml75_4165e1e3-7227-4666-a737-14083e64195d/nmstate-handler/0.log" Dec 04 15:01:35 crc kubenswrapper[4715]: I1204 15:01:35.630574 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-nc82h_c4ab73a9-a74b-43dc-bbe5-ad71120feab2/kube-rbac-proxy/0.log" Dec 04 15:01:35 crc kubenswrapper[4715]: I1204 15:01:35.713534 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-nc82h_c4ab73a9-a74b-43dc-bbe5-ad71120feab2/nmstate-metrics/0.log" Dec 04 15:01:35 crc kubenswrapper[4715]: I1204 15:01:35.834628 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-8k4vh_48b345c0-a5f8-4ff5-b696-e147e5ec9753/nmstate-operator/0.log" Dec 04 15:01:35 crc kubenswrapper[4715]: I1204 15:01:35.936886 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-dbfzg_746b0b9b-20ec-42d8-8d9a-4aa11c18a58e/nmstate-webhook/0.log" Dec 04 15:01:38 crc kubenswrapper[4715]: I1204 15:01:38.758581 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:01:38 crc kubenswrapper[4715]: I1204 15:01:38.759154 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:01:38 crc kubenswrapper[4715]: I1204 15:01:38.759207 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 15:01:38 crc kubenswrapper[4715]: I1204 15:01:38.760018 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:01:38 crc kubenswrapper[4715]: I1204 15:01:38.760110 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" gracePeriod=600 Dec 04 15:01:38 crc kubenswrapper[4715]: E1204 15:01:38.893692 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:01:39 crc kubenswrapper[4715]: I1204 15:01:39.427159 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" exitCode=0 Dec 04 15:01:39 crc kubenswrapper[4715]: I1204 15:01:39.427237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375"} Dec 04 15:01:39 crc kubenswrapper[4715]: I1204 15:01:39.427534 4715 scope.go:117] "RemoveContainer" containerID="e2e7792c7b342edbef8b148ec4e4a9cd95546575483ca8dead24c87263498a7c" Dec 04 15:01:39 crc kubenswrapper[4715]: I1204 15:01:39.428298 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:01:39 crc kubenswrapper[4715]: E1204 15:01:39.428643 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.435136 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-kp8gj_c4a52cd8-e319-4888-8920-a69547290700/kube-rbac-proxy/0.log" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.525375 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-kp8gj_c4a52cd8-e319-4888-8920-a69547290700/controller/0.log" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.650709 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.860186 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.875353 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.908027 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:01:49 crc kubenswrapper[4715]: I1204 15:01:49.924632 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.056745 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.077348 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.107880 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.151012 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.312257 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.312306 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.356778 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/controller/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.369799 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.506157 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/frr-metrics/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.559222 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/kube-rbac-proxy-frr/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.564750 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/kube-rbac-proxy/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.777625 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/reloader/0.log" Dec 04 15:01:50 crc kubenswrapper[4715]: I1204 15:01:50.842866 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-znkxb_d254ca37-56e1-470f-839b-cc07aa846cd3/frr-k8s-webhook-server/0.log" Dec 04 15:01:51 crc kubenswrapper[4715]: I1204 15:01:51.013221 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d7c69d77f-4r8gn_a57d59fd-4f87-4221-a861-8a4cde8314d6/manager/0.log" Dec 04 15:01:51 crc kubenswrapper[4715]: I1204 15:01:51.181024 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:01:51 crc kubenswrapper[4715]: E1204 15:01:51.181839 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:01:51 crc kubenswrapper[4715]: I1204 15:01:51.251855 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-78dd5c7fc8-vml6t_0346d3fa-78df-4769-9c41-b1b1ca9682b9/webhook-server/0.log" Dec 04 15:01:51 crc kubenswrapper[4715]: I1204 15:01:51.322909 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mx2g2_1de7de85-bcce-4132-9885-45d7996f6845/kube-rbac-proxy/0.log" Dec 04 15:01:51 crc kubenswrapper[4715]: I1204 15:01:51.986978 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mx2g2_1de7de85-bcce-4132-9885-45d7996f6845/speaker/0.log" Dec 04 15:01:52 crc kubenswrapper[4715]: I1204 15:01:52.144574 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/frr/0.log" Dec 04 15:02:03 crc kubenswrapper[4715]: I1204 15:02:03.797619 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/util/0.log" Dec 04 15:02:03 crc kubenswrapper[4715]: I1204 15:02:03.915832 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/pull/0.log" Dec 04 15:02:03 crc kubenswrapper[4715]: I1204 15:02:03.946106 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/util/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.021536 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/pull/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.199295 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:02:04 crc kubenswrapper[4715]: E1204 15:02:04.199557 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.369805 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/pull/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.403910 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/util/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.427793 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/extract/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.912791 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/util/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.912852 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/util/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.914630 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/pull/0.log" Dec 04 15:02:04 crc kubenswrapper[4715]: I1204 15:02:04.914723 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/pull/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.076914 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/util/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.077738 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/pull/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.140101 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/extract/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.275204 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-utilities/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.470269 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-utilities/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.497152 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-content/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.533155 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-content/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.680225 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-utilities/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.732682 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-content/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.796649 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/registry-server/0.log" Dec 04 15:02:05 crc kubenswrapper[4715]: I1204 15:02:05.917228 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-utilities/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.098993 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-utilities/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.102159 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-content/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.123595 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-content/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.321850 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-content/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.322216 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-utilities/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.620666 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-utilities/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.673782 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7wkp2_be7b5959-83d4-44ef-a596-f52b2e2a34f7/marketplace-operator/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.939067 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-content/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.944624 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-utilities/0.log" Dec 04 15:02:06 crc kubenswrapper[4715]: I1204 15:02:06.961399 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-content/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.036075 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/registry-server/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.137747 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-content/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.137750 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-utilities/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.274415 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/registry-server/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.357650 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-utilities/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.573350 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-content/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.573350 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-utilities/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.600915 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-content/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.797239 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-content/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.888869 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-utilities/0.log" Dec 04 15:02:07 crc kubenswrapper[4715]: I1204 15:02:07.915230 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/registry-server/0.log" Dec 04 15:02:18 crc kubenswrapper[4715]: I1204 15:02:18.181198 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:02:18 crc kubenswrapper[4715]: E1204 15:02:18.181958 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:02:33 crc kubenswrapper[4715]: I1204 15:02:33.183789 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:02:33 crc kubenswrapper[4715]: E1204 15:02:33.184520 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:02:45 crc kubenswrapper[4715]: I1204 15:02:45.618350 4715 scope.go:117] "RemoveContainer" containerID="d40aca24e416885673aa8f7ef3fb2ec764f6d4715303f4f0707b273c1cdd7c54" Dec 04 15:02:45 crc kubenswrapper[4715]: I1204 15:02:45.707627 4715 scope.go:117] "RemoveContainer" containerID="e71838da48d31258d6653e5bf2b0783c1fc9018c0afcff8674561fd6c1a6bbf7" Dec 04 15:02:45 crc kubenswrapper[4715]: I1204 15:02:45.726669 4715 scope.go:117] "RemoveContainer" containerID="1ce732253d8801cea9a33a0a4ff55e1f2cdd60c5d98416e37805d3d3b40ec46a" Dec 04 15:02:47 crc kubenswrapper[4715]: I1204 15:02:47.181370 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:02:47 crc kubenswrapper[4715]: E1204 15:02:47.182168 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:03:01 crc kubenswrapper[4715]: I1204 15:03:01.186617 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:03:01 crc kubenswrapper[4715]: E1204 15:03:01.187347 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:03:16 crc kubenswrapper[4715]: I1204 15:03:16.181364 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:03:16 crc kubenswrapper[4715]: E1204 15:03:16.182375 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:03:28 crc kubenswrapper[4715]: I1204 15:03:28.181603 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:03:28 crc kubenswrapper[4715]: E1204 15:03:28.183746 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:03:41 crc kubenswrapper[4715]: I1204 15:03:41.181008 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:03:41 crc kubenswrapper[4715]: E1204 15:03:41.182092 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:03:51 crc kubenswrapper[4715]: I1204 15:03:51.722837 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerID="49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865" exitCode=0 Dec 04 15:03:51 crc kubenswrapper[4715]: I1204 15:03:51.722934 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn5hf/must-gather-88prb" event={"ID":"ba37b5d6-3f67-4c36-bfd0-052e0989023a","Type":"ContainerDied","Data":"49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865"} Dec 04 15:03:51 crc kubenswrapper[4715]: I1204 15:03:51.724308 4715 scope.go:117] "RemoveContainer" containerID="49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865" Dec 04 15:03:51 crc kubenswrapper[4715]: I1204 15:03:51.951475 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zn5hf_must-gather-88prb_ba37b5d6-3f67-4c36-bfd0-052e0989023a/gather/0.log" Dec 04 15:03:54 crc kubenswrapper[4715]: I1204 15:03:54.182220 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:03:54 crc kubenswrapper[4715]: E1204 15:03:54.182998 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.065426 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn5hf/must-gather-88prb"] Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.066177 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zn5hf/must-gather-88prb" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="copy" containerID="cri-o://e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef" gracePeriod=2 Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.079049 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn5hf/must-gather-88prb"] Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.498298 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zn5hf_must-gather-88prb_ba37b5d6-3f67-4c36-bfd0-052e0989023a/copy/0.log" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.499341 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.602109 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbc7b\" (UniqueName: \"kubernetes.io/projected/ba37b5d6-3f67-4c36-bfd0-052e0989023a-kube-api-access-nbc7b\") pod \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.602590 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ba37b5d6-3f67-4c36-bfd0-052e0989023a-must-gather-output\") pod \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\" (UID: \"ba37b5d6-3f67-4c36-bfd0-052e0989023a\") " Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.620026 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba37b5d6-3f67-4c36-bfd0-052e0989023a-kube-api-access-nbc7b" (OuterVolumeSpecName: "kube-api-access-nbc7b") pod "ba37b5d6-3f67-4c36-bfd0-052e0989023a" (UID: "ba37b5d6-3f67-4c36-bfd0-052e0989023a"). InnerVolumeSpecName "kube-api-access-nbc7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.715805 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbc7b\" (UniqueName: \"kubernetes.io/projected/ba37b5d6-3f67-4c36-bfd0-052e0989023a-kube-api-access-nbc7b\") on node \"crc\" DevicePath \"\"" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.745123 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba37b5d6-3f67-4c36-bfd0-052e0989023a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ba37b5d6-3f67-4c36-bfd0-052e0989023a" (UID: "ba37b5d6-3f67-4c36-bfd0-052e0989023a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.799768 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zn5hf_must-gather-88prb_ba37b5d6-3f67-4c36-bfd0-052e0989023a/copy/0.log" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.800307 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerID="e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef" exitCode=143 Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.800352 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn5hf/must-gather-88prb" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.800405 4715 scope.go:117] "RemoveContainer" containerID="e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.817300 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ba37b5d6-3f67-4c36-bfd0-052e0989023a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.824214 4715 scope.go:117] "RemoveContainer" containerID="49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.885485 4715 scope.go:117] "RemoveContainer" containerID="e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef" Dec 04 15:04:00 crc kubenswrapper[4715]: E1204 15:04:00.886383 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef\": container with ID starting with e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef not found: ID does not exist" containerID="e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.886447 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef"} err="failed to get container status \"e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef\": rpc error: code = NotFound desc = could not find container \"e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef\": container with ID starting with e958f64638516054994970ea4d8f2eaf66350972168036d0fb0803e36f7e75ef not found: ID does not exist" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.886475 4715 scope.go:117] "RemoveContainer" containerID="49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865" Dec 04 15:04:00 crc kubenswrapper[4715]: E1204 15:04:00.886845 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865\": container with ID starting with 49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865 not found: ID does not exist" containerID="49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865" Dec 04 15:04:00 crc kubenswrapper[4715]: I1204 15:04:00.886894 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865"} err="failed to get container status \"49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865\": rpc error: code = NotFound desc = could not find container \"49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865\": container with ID starting with 49d870e630c2c256f6640799b215a32b7a60bbbdbd48255fe1e406b545ffb865 not found: ID does not exist" Dec 04 15:04:01 crc kubenswrapper[4715]: I1204 15:04:01.191619 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" path="/var/lib/kubelet/pods/ba37b5d6-3f67-4c36-bfd0-052e0989023a/volumes" Dec 04 15:04:05 crc kubenswrapper[4715]: I1204 15:04:05.180882 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:04:05 crc kubenswrapper[4715]: E1204 15:04:05.181727 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:04:17 crc kubenswrapper[4715]: I1204 15:04:17.182369 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:04:17 crc kubenswrapper[4715]: E1204 15:04:17.183356 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:04:30 crc kubenswrapper[4715]: I1204 15:04:30.181225 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:04:30 crc kubenswrapper[4715]: E1204 15:04:30.181899 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:04:41 crc kubenswrapper[4715]: I1204 15:04:41.180814 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:04:41 crc kubenswrapper[4715]: E1204 15:04:41.181652 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:04:53 crc kubenswrapper[4715]: I1204 15:04:53.189595 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:04:53 crc kubenswrapper[4715]: E1204 15:04:53.190749 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.881876 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bcmnz"] Dec 04 15:04:57 crc kubenswrapper[4715]: E1204 15:04:57.882975 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="gather" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.882995 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="gather" Dec 04 15:04:57 crc kubenswrapper[4715]: E1204 15:04:57.883026 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="copy" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.883056 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="copy" Dec 04 15:04:57 crc kubenswrapper[4715]: E1204 15:04:57.883081 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" containerName="keystone-cron" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.883093 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" containerName="keystone-cron" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.883337 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="copy" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.883366 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e0a90fb-c84b-45ee-ae9f-7975ff939ef0" containerName="keystone-cron" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.883378 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba37b5d6-3f67-4c36-bfd0-052e0989023a" containerName="gather" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.885239 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:57 crc kubenswrapper[4715]: I1204 15:04:57.893865 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcmnz"] Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.004902 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-utilities\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.005245 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmxwn\" (UniqueName: \"kubernetes.io/projected/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-kube-api-access-kmxwn\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.005442 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-catalog-content\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.107390 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-catalog-content\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.107517 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-utilities\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.107654 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmxwn\" (UniqueName: \"kubernetes.io/projected/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-kube-api-access-kmxwn\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.108058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-catalog-content\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.108103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-utilities\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.144862 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmxwn\" (UniqueName: \"kubernetes.io/projected/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-kube-api-access-kmxwn\") pod \"redhat-operators-bcmnz\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.205812 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:04:58 crc kubenswrapper[4715]: I1204 15:04:58.698741 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcmnz"] Dec 04 15:04:59 crc kubenswrapper[4715]: I1204 15:04:59.334659 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerID="07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96" exitCode=0 Dec 04 15:04:59 crc kubenswrapper[4715]: I1204 15:04:59.334998 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerDied","Data":"07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96"} Dec 04 15:04:59 crc kubenswrapper[4715]: I1204 15:04:59.335023 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerStarted","Data":"361de8c518df442f564ae57b69129fa05f8f1326d36936652da22c6d41bbec2b"} Dec 04 15:05:03 crc kubenswrapper[4715]: I1204 15:05:03.394238 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerStarted","Data":"e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537"} Dec 04 15:05:05 crc kubenswrapper[4715]: I1204 15:05:05.182781 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:05:05 crc kubenswrapper[4715]: E1204 15:05:05.183372 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:05:09 crc kubenswrapper[4715]: I1204 15:05:09.445462 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerID="e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537" exitCode=0 Dec 04 15:05:09 crc kubenswrapper[4715]: I1204 15:05:09.445508 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerDied","Data":"e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537"} Dec 04 15:05:11 crc kubenswrapper[4715]: I1204 15:05:11.467916 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerStarted","Data":"eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c"} Dec 04 15:05:11 crc kubenswrapper[4715]: I1204 15:05:11.491996 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bcmnz" podStartSLOduration=4.150328171 podStartE2EDuration="14.491971057s" podCreationTimestamp="2025-12-04 15:04:57 +0000 UTC" firstStartedPulling="2025-12-04 15:05:00.345996126 +0000 UTC m=+4057.414714341" lastFinishedPulling="2025-12-04 15:05:10.687639012 +0000 UTC m=+4067.756357227" observedRunningTime="2025-12-04 15:05:11.485623853 +0000 UTC m=+4068.554342088" watchObservedRunningTime="2025-12-04 15:05:11.491971057 +0000 UTC m=+4068.560689272" Dec 04 15:05:18 crc kubenswrapper[4715]: I1204 15:05:18.182853 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:05:18 crc kubenswrapper[4715]: E1204 15:05:18.183735 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:05:18 crc kubenswrapper[4715]: I1204 15:05:18.207350 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:05:18 crc kubenswrapper[4715]: I1204 15:05:18.207425 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:05:18 crc kubenswrapper[4715]: I1204 15:05:18.261759 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:05:18 crc kubenswrapper[4715]: I1204 15:05:18.810890 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:05:18 crc kubenswrapper[4715]: I1204 15:05:18.862955 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcmnz"] Dec 04 15:05:20 crc kubenswrapper[4715]: I1204 15:05:20.548058 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bcmnz" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="registry-server" containerID="cri-o://eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c" gracePeriod=2 Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.146705 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.264452 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-utilities\") pod \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.264961 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmxwn\" (UniqueName: \"kubernetes.io/projected/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-kube-api-access-kmxwn\") pod \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.264994 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-catalog-content\") pod \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\" (UID: \"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e\") " Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.265469 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-utilities" (OuterVolumeSpecName: "utilities") pod "7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" (UID: "7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.265762 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.271392 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-kube-api-access-kmxwn" (OuterVolumeSpecName: "kube-api-access-kmxwn") pod "7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" (UID: "7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e"). InnerVolumeSpecName "kube-api-access-kmxwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.368223 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmxwn\" (UniqueName: \"kubernetes.io/projected/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-kube-api-access-kmxwn\") on node \"crc\" DevicePath \"\"" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.379152 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" (UID: "7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.469897 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.570371 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerID="eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c" exitCode=0 Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.570477 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcmnz" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.571578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerDied","Data":"eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c"} Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.571700 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcmnz" event={"ID":"7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e","Type":"ContainerDied","Data":"361de8c518df442f564ae57b69129fa05f8f1326d36936652da22c6d41bbec2b"} Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.571792 4715 scope.go:117] "RemoveContainer" containerID="eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.599507 4715 scope.go:117] "RemoveContainer" containerID="e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.617173 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcmnz"] Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.626916 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bcmnz"] Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.630351 4715 scope.go:117] "RemoveContainer" containerID="07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.684781 4715 scope.go:117] "RemoveContainer" containerID="eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c" Dec 04 15:05:21 crc kubenswrapper[4715]: E1204 15:05:21.685821 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c\": container with ID starting with eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c not found: ID does not exist" containerID="eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.685881 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c"} err="failed to get container status \"eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c\": rpc error: code = NotFound desc = could not find container \"eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c\": container with ID starting with eb125f8d73d8b8a00e9a80bc9dd7612980ba7a802c31010d09aecd94d1c4df7c not found: ID does not exist" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.685914 4715 scope.go:117] "RemoveContainer" containerID="e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537" Dec 04 15:05:21 crc kubenswrapper[4715]: E1204 15:05:21.686301 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537\": container with ID starting with e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537 not found: ID does not exist" containerID="e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.686332 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537"} err="failed to get container status \"e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537\": rpc error: code = NotFound desc = could not find container \"e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537\": container with ID starting with e04236e4b033f4c2b47a8427a8230ecf301970aad659233b3cbb8cd81bd29537 not found: ID does not exist" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.686354 4715 scope.go:117] "RemoveContainer" containerID="07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96" Dec 04 15:05:21 crc kubenswrapper[4715]: E1204 15:05:21.686888 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96\": container with ID starting with 07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96 not found: ID does not exist" containerID="07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96" Dec 04 15:05:21 crc kubenswrapper[4715]: I1204 15:05:21.686953 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96"} err="failed to get container status \"07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96\": rpc error: code = NotFound desc = could not find container \"07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96\": container with ID starting with 07afe1bcfd24073d6d80118c40715b9ab09d07c112236d6aa76aca0319b30d96 not found: ID does not exist" Dec 04 15:05:23 crc kubenswrapper[4715]: I1204 15:05:23.209387 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" path="/var/lib/kubelet/pods/7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e/volumes" Dec 04 15:05:31 crc kubenswrapper[4715]: I1204 15:05:31.180781 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:05:31 crc kubenswrapper[4715]: E1204 15:05:31.181772 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:05:43 crc kubenswrapper[4715]: I1204 15:05:43.188432 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:05:43 crc kubenswrapper[4715]: E1204 15:05:43.189479 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:05:45 crc kubenswrapper[4715]: I1204 15:05:45.825170 4715 scope.go:117] "RemoveContainer" containerID="c8da63f77942de22144be60f5f2889044db8eaca19f6d424edba997f9d85a6c7" Dec 04 15:05:54 crc kubenswrapper[4715]: I1204 15:05:54.180490 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:05:54 crc kubenswrapper[4715]: E1204 15:05:54.181301 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:06:05 crc kubenswrapper[4715]: I1204 15:06:05.183728 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:06:05 crc kubenswrapper[4715]: E1204 15:06:05.184575 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:06:20 crc kubenswrapper[4715]: I1204 15:06:20.181141 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:06:20 crc kubenswrapper[4715]: E1204 15:06:20.182264 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:06:35 crc kubenswrapper[4715]: I1204 15:06:35.181476 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:06:35 crc kubenswrapper[4715]: E1204 15:06:35.182398 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.181664 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.622381 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-54mfk"] Dec 04 15:06:47 crc kubenswrapper[4715]: E1204 15:06:47.622904 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="registry-server" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.622925 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="registry-server" Dec 04 15:06:47 crc kubenswrapper[4715]: E1204 15:06:47.622969 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="extract-utilities" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.622978 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="extract-utilities" Dec 04 15:06:47 crc kubenswrapper[4715]: E1204 15:06:47.623005 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="extract-content" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.623013 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="extract-content" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.623274 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0ab3a9-c5b6-4d6f-bc4e-b8b9a179304e" containerName="registry-server" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.625025 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.632628 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-54mfk"] Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.665109 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gvgd\" (UniqueName: \"kubernetes.io/projected/e7dd4a92-a7c3-4e9b-a171-346303b23bca-kube-api-access-9gvgd\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.665198 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-catalog-content\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.665227 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-utilities\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.767438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gvgd\" (UniqueName: \"kubernetes.io/projected/e7dd4a92-a7c3-4e9b-a171-346303b23bca-kube-api-access-9gvgd\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.767536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-catalog-content\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.767562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-utilities\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.768183 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-utilities\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.768424 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-catalog-content\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.795814 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gvgd\" (UniqueName: \"kubernetes.io/projected/e7dd4a92-a7c3-4e9b-a171-346303b23bca-kube-api-access-9gvgd\") pod \"community-operators-54mfk\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:47 crc kubenswrapper[4715]: I1204 15:06:47.940888 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:48 crc kubenswrapper[4715]: I1204 15:06:48.396245 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"288ff8ad632dc3f0edd464ac0c1fad0e3b8cc2f3f601075dd1429fe6530ce81f"} Dec 04 15:06:48 crc kubenswrapper[4715]: I1204 15:06:48.530208 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-54mfk"] Dec 04 15:06:48 crc kubenswrapper[4715]: W1204 15:06:48.567207 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7dd4a92_a7c3_4e9b_a171_346303b23bca.slice/crio-81bc476543925e833ccdbbea2ad380c3c2250505e4f1f8f16c51633691cf6c53 WatchSource:0}: Error finding container 81bc476543925e833ccdbbea2ad380c3c2250505e4f1f8f16c51633691cf6c53: Status 404 returned error can't find the container with id 81bc476543925e833ccdbbea2ad380c3c2250505e4f1f8f16c51633691cf6c53 Dec 04 15:06:49 crc kubenswrapper[4715]: I1204 15:06:49.406511 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerID="879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02" exitCode=0 Dec 04 15:06:49 crc kubenswrapper[4715]: I1204 15:06:49.406680 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerDied","Data":"879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02"} Dec 04 15:06:49 crc kubenswrapper[4715]: I1204 15:06:49.406860 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerStarted","Data":"81bc476543925e833ccdbbea2ad380c3c2250505e4f1f8f16c51633691cf6c53"} Dec 04 15:06:49 crc kubenswrapper[4715]: I1204 15:06:49.409006 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:06:50 crc kubenswrapper[4715]: I1204 15:06:50.418741 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerStarted","Data":"ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d"} Dec 04 15:06:51 crc kubenswrapper[4715]: I1204 15:06:51.429731 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerID="ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d" exitCode=0 Dec 04 15:06:51 crc kubenswrapper[4715]: I1204 15:06:51.429829 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerDied","Data":"ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d"} Dec 04 15:06:52 crc kubenswrapper[4715]: I1204 15:06:52.452617 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerStarted","Data":"47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9"} Dec 04 15:06:52 crc kubenswrapper[4715]: I1204 15:06:52.483318 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-54mfk" podStartSLOduration=3.037988859 podStartE2EDuration="5.483290804s" podCreationTimestamp="2025-12-04 15:06:47 +0000 UTC" firstStartedPulling="2025-12-04 15:06:49.408660247 +0000 UTC m=+4166.477378462" lastFinishedPulling="2025-12-04 15:06:51.853962192 +0000 UTC m=+4168.922680407" observedRunningTime="2025-12-04 15:06:52.47469415 +0000 UTC m=+4169.543412375" watchObservedRunningTime="2025-12-04 15:06:52.483290804 +0000 UTC m=+4169.552009029" Dec 04 15:06:57 crc kubenswrapper[4715]: I1204 15:06:57.941068 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:57 crc kubenswrapper[4715]: I1204 15:06:57.941579 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:57 crc kubenswrapper[4715]: I1204 15:06:57.985911 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:58 crc kubenswrapper[4715]: I1204 15:06:58.557911 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:06:58 crc kubenswrapper[4715]: I1204 15:06:58.615351 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-54mfk"] Dec 04 15:07:00 crc kubenswrapper[4715]: I1204 15:07:00.523129 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-54mfk" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="registry-server" containerID="cri-o://47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9" gracePeriod=2 Dec 04 15:07:00 crc kubenswrapper[4715]: I1204 15:07:00.930674 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.019992 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-utilities\") pod \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.020468 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gvgd\" (UniqueName: \"kubernetes.io/projected/e7dd4a92-a7c3-4e9b-a171-346303b23bca-kube-api-access-9gvgd\") pod \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.020629 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-catalog-content\") pod \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\" (UID: \"e7dd4a92-a7c3-4e9b-a171-346303b23bca\") " Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.020656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-utilities" (OuterVolumeSpecName: "utilities") pod "e7dd4a92-a7c3-4e9b-a171-346303b23bca" (UID: "e7dd4a92-a7c3-4e9b-a171-346303b23bca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.021165 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.027293 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7dd4a92-a7c3-4e9b-a171-346303b23bca-kube-api-access-9gvgd" (OuterVolumeSpecName: "kube-api-access-9gvgd") pod "e7dd4a92-a7c3-4e9b-a171-346303b23bca" (UID: "e7dd4a92-a7c3-4e9b-a171-346303b23bca"). InnerVolumeSpecName "kube-api-access-9gvgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.070894 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7dd4a92-a7c3-4e9b-a171-346303b23bca" (UID: "e7dd4a92-a7c3-4e9b-a171-346303b23bca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.122720 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7dd4a92-a7c3-4e9b-a171-346303b23bca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.122764 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gvgd\" (UniqueName: \"kubernetes.io/projected/e7dd4a92-a7c3-4e9b-a171-346303b23bca-kube-api-access-9gvgd\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.536958 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerID="47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9" exitCode=0 Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.537014 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerDied","Data":"47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9"} Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.537069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-54mfk" event={"ID":"e7dd4a92-a7c3-4e9b-a171-346303b23bca","Type":"ContainerDied","Data":"81bc476543925e833ccdbbea2ad380c3c2250505e4f1f8f16c51633691cf6c53"} Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.537093 4715 scope.go:117] "RemoveContainer" containerID="47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.538213 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-54mfk" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.561946 4715 scope.go:117] "RemoveContainer" containerID="ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.572495 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-54mfk"] Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.583693 4715 scope.go:117] "RemoveContainer" containerID="879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.584439 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-54mfk"] Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.648836 4715 scope.go:117] "RemoveContainer" containerID="47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9" Dec 04 15:07:01 crc kubenswrapper[4715]: E1204 15:07:01.649459 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9\": container with ID starting with 47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9 not found: ID does not exist" containerID="47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.649528 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9"} err="failed to get container status \"47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9\": rpc error: code = NotFound desc = could not find container \"47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9\": container with ID starting with 47bb92862c7bf48878d40eb1e7fb64071298334a58d47b31689c5345ab54e3e9 not found: ID does not exist" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.649552 4715 scope.go:117] "RemoveContainer" containerID="ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d" Dec 04 15:07:01 crc kubenswrapper[4715]: E1204 15:07:01.649870 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d\": container with ID starting with ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d not found: ID does not exist" containerID="ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.649906 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d"} err="failed to get container status \"ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d\": rpc error: code = NotFound desc = could not find container \"ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d\": container with ID starting with ef82233412ee88a764a361029ec301caed07d09ee63799cf647ae99b4a84721d not found: ID does not exist" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.649924 4715 scope.go:117] "RemoveContainer" containerID="879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02" Dec 04 15:07:01 crc kubenswrapper[4715]: E1204 15:07:01.650306 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02\": container with ID starting with 879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02 not found: ID does not exist" containerID="879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02" Dec 04 15:07:01 crc kubenswrapper[4715]: I1204 15:07:01.650333 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02"} err="failed to get container status \"879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02\": rpc error: code = NotFound desc = could not find container \"879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02\": container with ID starting with 879d85ae7d732675bd0d86b8726d2fc72323e17afcbd9b665bc9eb0e7ef17a02 not found: ID does not exist" Dec 04 15:07:03 crc kubenswrapper[4715]: I1204 15:07:03.199327 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" path="/var/lib/kubelet/pods/e7dd4a92-a7c3-4e9b-a171-346303b23bca/volumes" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.498992 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7z8qr/must-gather-qgp2k"] Dec 04 15:07:09 crc kubenswrapper[4715]: E1204 15:07:09.499962 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="extract-utilities" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.499980 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="extract-utilities" Dec 04 15:07:09 crc kubenswrapper[4715]: E1204 15:07:09.500019 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="registry-server" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.500029 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="registry-server" Dec 04 15:07:09 crc kubenswrapper[4715]: E1204 15:07:09.500069 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="extract-content" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.500078 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="extract-content" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.500340 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7dd4a92-a7c3-4e9b-a171-346303b23bca" containerName="registry-server" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.502594 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.508493 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7z8qr"/"openshift-service-ca.crt" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.508649 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7z8qr"/"kube-root-ca.crt" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.519493 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7z8qr/must-gather-qgp2k"] Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.593310 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxwjl\" (UniqueName: \"kubernetes.io/projected/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-kube-api-access-hxwjl\") pod \"must-gather-qgp2k\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.593473 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-must-gather-output\") pod \"must-gather-qgp2k\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.696086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-must-gather-output\") pod \"must-gather-qgp2k\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.696234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxwjl\" (UniqueName: \"kubernetes.io/projected/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-kube-api-access-hxwjl\") pod \"must-gather-qgp2k\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.697108 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-must-gather-output\") pod \"must-gather-qgp2k\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.715652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxwjl\" (UniqueName: \"kubernetes.io/projected/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-kube-api-access-hxwjl\") pod \"must-gather-qgp2k\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:09 crc kubenswrapper[4715]: I1204 15:07:09.869027 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:07:10 crc kubenswrapper[4715]: I1204 15:07:10.354194 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7z8qr/must-gather-qgp2k"] Dec 04 15:07:10 crc kubenswrapper[4715]: W1204 15:07:10.363084 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc242cfd_fb20_49c1_8812_bc13d5ccb2f8.slice/crio-684874aaff63ba3861cc412f1dc3340189bef4e3598e246f4b3a02cdd38ac292 WatchSource:0}: Error finding container 684874aaff63ba3861cc412f1dc3340189bef4e3598e246f4b3a02cdd38ac292: Status 404 returned error can't find the container with id 684874aaff63ba3861cc412f1dc3340189bef4e3598e246f4b3a02cdd38ac292 Dec 04 15:07:10 crc kubenswrapper[4715]: I1204 15:07:10.619663 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" event={"ID":"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8","Type":"ContainerStarted","Data":"684874aaff63ba3861cc412f1dc3340189bef4e3598e246f4b3a02cdd38ac292"} Dec 04 15:07:11 crc kubenswrapper[4715]: I1204 15:07:11.631161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" event={"ID":"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8","Type":"ContainerStarted","Data":"e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d"} Dec 04 15:07:11 crc kubenswrapper[4715]: I1204 15:07:11.631524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" event={"ID":"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8","Type":"ContainerStarted","Data":"005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d"} Dec 04 15:07:11 crc kubenswrapper[4715]: I1204 15:07:11.651138 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" podStartSLOduration=2.651116052 podStartE2EDuration="2.651116052s" podCreationTimestamp="2025-12-04 15:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:07:11.643811762 +0000 UTC m=+4188.712529987" watchObservedRunningTime="2025-12-04 15:07:11.651116052 +0000 UTC m=+4188.719834267" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.141024 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-h2wxg"] Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.143288 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.146462 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7z8qr"/"default-dockercfg-h2pz6" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.290850 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a60f018-ea5e-4b4f-aadf-e5a68812c983-host\") pod \"crc-debug-h2wxg\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.290950 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hs9m\" (UniqueName: \"kubernetes.io/projected/3a60f018-ea5e-4b4f-aadf-e5a68812c983-kube-api-access-2hs9m\") pod \"crc-debug-h2wxg\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.393257 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a60f018-ea5e-4b4f-aadf-e5a68812c983-host\") pod \"crc-debug-h2wxg\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.393315 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hs9m\" (UniqueName: \"kubernetes.io/projected/3a60f018-ea5e-4b4f-aadf-e5a68812c983-kube-api-access-2hs9m\") pod \"crc-debug-h2wxg\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.393577 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a60f018-ea5e-4b4f-aadf-e5a68812c983-host\") pod \"crc-debug-h2wxg\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.419121 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hs9m\" (UniqueName: \"kubernetes.io/projected/3a60f018-ea5e-4b4f-aadf-e5a68812c983-kube-api-access-2hs9m\") pod \"crc-debug-h2wxg\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.469981 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:14 crc kubenswrapper[4715]: I1204 15:07:14.657126 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" event={"ID":"3a60f018-ea5e-4b4f-aadf-e5a68812c983","Type":"ContainerStarted","Data":"feec2a96af797ad6cf59072c11305e8585ef5024e29cbb3517877e7acf7bfc7a"} Dec 04 15:07:15 crc kubenswrapper[4715]: I1204 15:07:15.691191 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" event={"ID":"3a60f018-ea5e-4b4f-aadf-e5a68812c983","Type":"ContainerStarted","Data":"90259b4a434ce2bee343c99e9bef04eb9c643d717a2aac29b6f3c9655d8e3d6f"} Dec 04 15:07:15 crc kubenswrapper[4715]: I1204 15:07:15.708357 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" podStartSLOduration=1.708341365 podStartE2EDuration="1.708341365s" podCreationTimestamp="2025-12-04 15:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:07:15.706948799 +0000 UTC m=+4192.775667014" watchObservedRunningTime="2025-12-04 15:07:15.708341365 +0000 UTC m=+4192.777059580" Dec 04 15:07:47 crc kubenswrapper[4715]: E1204 15:07:47.351568 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a60f018_ea5e_4b4f_aadf_e5a68812c983.slice/crio-conmon-90259b4a434ce2bee343c99e9bef04eb9c643d717a2aac29b6f3c9655d8e3d6f.scope\": RecentStats: unable to find data in memory cache]" Dec 04 15:07:47 crc kubenswrapper[4715]: I1204 15:07:47.984981 4715 generic.go:334] "Generic (PLEG): container finished" podID="3a60f018-ea5e-4b4f-aadf-e5a68812c983" containerID="90259b4a434ce2bee343c99e9bef04eb9c643d717a2aac29b6f3c9655d8e3d6f" exitCode=0 Dec 04 15:07:47 crc kubenswrapper[4715]: I1204 15:07:47.985087 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" event={"ID":"3a60f018-ea5e-4b4f-aadf-e5a68812c983","Type":"ContainerDied","Data":"90259b4a434ce2bee343c99e9bef04eb9c643d717a2aac29b6f3c9655d8e3d6f"} Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.116405 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.156282 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-h2wxg"] Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.166219 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-h2wxg"] Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.268691 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hs9m\" (UniqueName: \"kubernetes.io/projected/3a60f018-ea5e-4b4f-aadf-e5a68812c983-kube-api-access-2hs9m\") pod \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.268939 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a60f018-ea5e-4b4f-aadf-e5a68812c983-host\") pod \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\" (UID: \"3a60f018-ea5e-4b4f-aadf-e5a68812c983\") " Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.270394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a60f018-ea5e-4b4f-aadf-e5a68812c983-host" (OuterVolumeSpecName: "host") pod "3a60f018-ea5e-4b4f-aadf-e5a68812c983" (UID: "3a60f018-ea5e-4b4f-aadf-e5a68812c983"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.275976 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a60f018-ea5e-4b4f-aadf-e5a68812c983-kube-api-access-2hs9m" (OuterVolumeSpecName: "kube-api-access-2hs9m") pod "3a60f018-ea5e-4b4f-aadf-e5a68812c983" (UID: "3a60f018-ea5e-4b4f-aadf-e5a68812c983"). InnerVolumeSpecName "kube-api-access-2hs9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.371102 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hs9m\" (UniqueName: \"kubernetes.io/projected/3a60f018-ea5e-4b4f-aadf-e5a68812c983-kube-api-access-2hs9m\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:49 crc kubenswrapper[4715]: I1204 15:07:49.371139 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a60f018-ea5e-4b4f-aadf-e5a68812c983-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.004001 4715 scope.go:117] "RemoveContainer" containerID="90259b4a434ce2bee343c99e9bef04eb9c643d717a2aac29b6f3c9655d8e3d6f" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.004166 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-h2wxg" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.325842 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-spp7z"] Dec 04 15:07:50 crc kubenswrapper[4715]: E1204 15:07:50.326632 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a60f018-ea5e-4b4f-aadf-e5a68812c983" containerName="container-00" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.326647 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a60f018-ea5e-4b4f-aadf-e5a68812c983" containerName="container-00" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.326915 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a60f018-ea5e-4b4f-aadf-e5a68812c983" containerName="container-00" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.327516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.329838 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7z8qr"/"default-dockercfg-h2pz6" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.490516 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5944a00f-a7da-484e-9aba-999cd67f3146-host\") pod \"crc-debug-spp7z\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.490882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbz8\" (UniqueName: \"kubernetes.io/projected/5944a00f-a7da-484e-9aba-999cd67f3146-kube-api-access-mfbz8\") pod \"crc-debug-spp7z\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.592771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbz8\" (UniqueName: \"kubernetes.io/projected/5944a00f-a7da-484e-9aba-999cd67f3146-kube-api-access-mfbz8\") pod \"crc-debug-spp7z\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.592957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5944a00f-a7da-484e-9aba-999cd67f3146-host\") pod \"crc-debug-spp7z\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.593171 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5944a00f-a7da-484e-9aba-999cd67f3146-host\") pod \"crc-debug-spp7z\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.615386 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbz8\" (UniqueName: \"kubernetes.io/projected/5944a00f-a7da-484e-9aba-999cd67f3146-kube-api-access-mfbz8\") pod \"crc-debug-spp7z\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:50 crc kubenswrapper[4715]: I1204 15:07:50.642064 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:51 crc kubenswrapper[4715]: I1204 15:07:51.013897 4715 generic.go:334] "Generic (PLEG): container finished" podID="5944a00f-a7da-484e-9aba-999cd67f3146" containerID="5335f11eaf05cc58a4f7af51f080de60183870a634d96d8f161210d64cbeb785" exitCode=0 Dec 04 15:07:51 crc kubenswrapper[4715]: I1204 15:07:51.013986 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-spp7z" event={"ID":"5944a00f-a7da-484e-9aba-999cd67f3146","Type":"ContainerDied","Data":"5335f11eaf05cc58a4f7af51f080de60183870a634d96d8f161210d64cbeb785"} Dec 04 15:07:51 crc kubenswrapper[4715]: I1204 15:07:51.014560 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-spp7z" event={"ID":"5944a00f-a7da-484e-9aba-999cd67f3146","Type":"ContainerStarted","Data":"92b6a3f287436f1f4007e09e8c52e6a8fda7f4a69e5f7c9b82809e1898c6a538"} Dec 04 15:07:51 crc kubenswrapper[4715]: I1204 15:07:51.195277 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a60f018-ea5e-4b4f-aadf-e5a68812c983" path="/var/lib/kubelet/pods/3a60f018-ea5e-4b4f-aadf-e5a68812c983/volumes" Dec 04 15:07:51 crc kubenswrapper[4715]: I1204 15:07:51.567912 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-spp7z"] Dec 04 15:07:51 crc kubenswrapper[4715]: I1204 15:07:51.577815 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-spp7z"] Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.124575 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.218565 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbz8\" (UniqueName: \"kubernetes.io/projected/5944a00f-a7da-484e-9aba-999cd67f3146-kube-api-access-mfbz8\") pod \"5944a00f-a7da-484e-9aba-999cd67f3146\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.218644 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5944a00f-a7da-484e-9aba-999cd67f3146-host\") pod \"5944a00f-a7da-484e-9aba-999cd67f3146\" (UID: \"5944a00f-a7da-484e-9aba-999cd67f3146\") " Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.218967 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5944a00f-a7da-484e-9aba-999cd67f3146-host" (OuterVolumeSpecName: "host") pod "5944a00f-a7da-484e-9aba-999cd67f3146" (UID: "5944a00f-a7da-484e-9aba-999cd67f3146"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.219290 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5944a00f-a7da-484e-9aba-999cd67f3146-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.234285 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5944a00f-a7da-484e-9aba-999cd67f3146-kube-api-access-mfbz8" (OuterVolumeSpecName: "kube-api-access-mfbz8") pod "5944a00f-a7da-484e-9aba-999cd67f3146" (UID: "5944a00f-a7da-484e-9aba-999cd67f3146"). InnerVolumeSpecName "kube-api-access-mfbz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.321014 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbz8\" (UniqueName: \"kubernetes.io/projected/5944a00f-a7da-484e-9aba-999cd67f3146-kube-api-access-mfbz8\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.719603 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-26ph9"] Dec 04 15:07:52 crc kubenswrapper[4715]: E1204 15:07:52.719993 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5944a00f-a7da-484e-9aba-999cd67f3146" containerName="container-00" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.720006 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5944a00f-a7da-484e-9aba-999cd67f3146" containerName="container-00" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.720239 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5944a00f-a7da-484e-9aba-999cd67f3146" containerName="container-00" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.720808 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.829890 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd42z\" (UniqueName: \"kubernetes.io/projected/493fc4e9-b874-4f0a-a14e-c346b482cfa5-kube-api-access-cd42z\") pod \"crc-debug-26ph9\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.829966 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/493fc4e9-b874-4f0a-a14e-c346b482cfa5-host\") pod \"crc-debug-26ph9\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.931743 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd42z\" (UniqueName: \"kubernetes.io/projected/493fc4e9-b874-4f0a-a14e-c346b482cfa5-kube-api-access-cd42z\") pod \"crc-debug-26ph9\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.932145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/493fc4e9-b874-4f0a-a14e-c346b482cfa5-host\") pod \"crc-debug-26ph9\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.932305 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/493fc4e9-b874-4f0a-a14e-c346b482cfa5-host\") pod \"crc-debug-26ph9\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:52 crc kubenswrapper[4715]: I1204 15:07:52.950623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd42z\" (UniqueName: \"kubernetes.io/projected/493fc4e9-b874-4f0a-a14e-c346b482cfa5-kube-api-access-cd42z\") pod \"crc-debug-26ph9\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:53 crc kubenswrapper[4715]: I1204 15:07:53.036787 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:53 crc kubenswrapper[4715]: I1204 15:07:53.048670 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92b6a3f287436f1f4007e09e8c52e6a8fda7f4a69e5f7c9b82809e1898c6a538" Dec 04 15:07:53 crc kubenswrapper[4715]: I1204 15:07:53.048758 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-spp7z" Dec 04 15:07:53 crc kubenswrapper[4715]: I1204 15:07:53.193242 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5944a00f-a7da-484e-9aba-999cd67f3146" path="/var/lib/kubelet/pods/5944a00f-a7da-484e-9aba-999cd67f3146/volumes" Dec 04 15:07:54 crc kubenswrapper[4715]: I1204 15:07:54.061901 4715 generic.go:334] "Generic (PLEG): container finished" podID="493fc4e9-b874-4f0a-a14e-c346b482cfa5" containerID="4f6596d2824ec8ccce643ef0ebf846a1318145c5cef511cd07ffb2c9f0cb7c0e" exitCode=0 Dec 04 15:07:54 crc kubenswrapper[4715]: I1204 15:07:54.062003 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-26ph9" event={"ID":"493fc4e9-b874-4f0a-a14e-c346b482cfa5","Type":"ContainerDied","Data":"4f6596d2824ec8ccce643ef0ebf846a1318145c5cef511cd07ffb2c9f0cb7c0e"} Dec 04 15:07:54 crc kubenswrapper[4715]: I1204 15:07:54.062269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/crc-debug-26ph9" event={"ID":"493fc4e9-b874-4f0a-a14e-c346b482cfa5","Type":"ContainerStarted","Data":"1fce9f00f4e4bee0fcd73ee2a368530934f832520a2ab90f5214f4c675fc345e"} Dec 04 15:07:54 crc kubenswrapper[4715]: I1204 15:07:54.106343 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-26ph9"] Dec 04 15:07:54 crc kubenswrapper[4715]: I1204 15:07:54.115439 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7z8qr/crc-debug-26ph9"] Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.182240 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.275802 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/493fc4e9-b874-4f0a-a14e-c346b482cfa5-host\") pod \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.275902 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd42z\" (UniqueName: \"kubernetes.io/projected/493fc4e9-b874-4f0a-a14e-c346b482cfa5-kube-api-access-cd42z\") pod \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\" (UID: \"493fc4e9-b874-4f0a-a14e-c346b482cfa5\") " Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.275924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/493fc4e9-b874-4f0a-a14e-c346b482cfa5-host" (OuterVolumeSpecName: "host") pod "493fc4e9-b874-4f0a-a14e-c346b482cfa5" (UID: "493fc4e9-b874-4f0a-a14e-c346b482cfa5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.276394 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/493fc4e9-b874-4f0a-a14e-c346b482cfa5-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.284497 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493fc4e9-b874-4f0a-a14e-c346b482cfa5-kube-api-access-cd42z" (OuterVolumeSpecName: "kube-api-access-cd42z") pod "493fc4e9-b874-4f0a-a14e-c346b482cfa5" (UID: "493fc4e9-b874-4f0a-a14e-c346b482cfa5"). InnerVolumeSpecName "kube-api-access-cd42z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:07:55 crc kubenswrapper[4715]: I1204 15:07:55.379823 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd42z\" (UniqueName: \"kubernetes.io/projected/493fc4e9-b874-4f0a-a14e-c346b482cfa5-kube-api-access-cd42z\") on node \"crc\" DevicePath \"\"" Dec 04 15:07:56 crc kubenswrapper[4715]: I1204 15:07:56.079648 4715 scope.go:117] "RemoveContainer" containerID="4f6596d2824ec8ccce643ef0ebf846a1318145c5cef511cd07ffb2c9f0cb7c0e" Dec 04 15:07:56 crc kubenswrapper[4715]: I1204 15:07:56.080026 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/crc-debug-26ph9" Dec 04 15:07:57 crc kubenswrapper[4715]: I1204 15:07:57.191857 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493fc4e9-b874-4f0a-a14e-c346b482cfa5" path="/var/lib/kubelet/pods/493fc4e9-b874-4f0a-a14e-c346b482cfa5/volumes" Dec 04 15:08:16 crc kubenswrapper[4715]: I1204 15:08:16.866572 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65b4dd5988-fzzwk_795cce68-d6db-4923-ab86-c474d5085318/barbican-api/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.027360 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65b4dd5988-fzzwk_795cce68-d6db-4923-ab86-c474d5085318/barbican-api-log/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.076132 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f7449dddd-lkzvw_589e9c2f-67f8-4208-aaa3-b391d59b7d01/barbican-keystone-listener/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.105256 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f7449dddd-lkzvw_589e9c2f-67f8-4208-aaa3-b391d59b7d01/barbican-keystone-listener-log/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.235498 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d967d55f-xm4d6_5a34d667-48aa-4bb9-82fa-bb528e9bcc20/barbican-worker/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.262196 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d967d55f-xm4d6_5a34d667-48aa-4bb9-82fa-bb528e9bcc20/barbican-worker-log/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.453822 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wsjjt_b6a57886-292b-4dac-bfb6-a51971fcde74/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.520616 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/ceilometer-central-agent/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.559561 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/ceilometer-notification-agent/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.651382 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/proxy-httpd/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.668310 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_347af36d-eea9-4d43-96e3-c20d6915ff4d/sg-core/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.791839 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_87853446-d88c-4d46-9fea-ff90af5617ba/cinder-api/0.log" Dec 04 15:08:17 crc kubenswrapper[4715]: I1204 15:08:17.878528 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_87853446-d88c-4d46-9fea-ff90af5617ba/cinder-api-log/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.003638 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6816724d-3e76-4832-9e96-ab6ef342c354/cinder-scheduler/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.044231 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6816724d-3e76-4832-9e96-ab6ef342c354/probe/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.309610 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6t47w_7f4cab97-fa52-47c4-ad8e-7469556526cc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.348059 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-t4kkq_337aa155-3f3f-4e84-a350-29d27ef13260/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.561748 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-rbrt6_dfe175eb-a594-4a23-a057-b3cbc53cad3e/init/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.658753 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-rbrt6_dfe175eb-a594-4a23-a057-b3cbc53cad3e/init/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.721342 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-rbrt6_dfe175eb-a594-4a23-a057-b3cbc53cad3e/dnsmasq-dns/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.820898 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-8dv88_a8a8da4d-f578-4428-8b74-9d4efd5f78ee/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.974061 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92c1ce5a-911d-4567-8d23-11777a6df443/glance-httpd/0.log" Dec 04 15:08:18 crc kubenswrapper[4715]: I1204 15:08:18.995416 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92c1ce5a-911d-4567-8d23-11777a6df443/glance-log/0.log" Dec 04 15:08:19 crc kubenswrapper[4715]: I1204 15:08:19.163188 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3bea4af7-30c3-4278-99d7-01f1c604a89c/glance-httpd/0.log" Dec 04 15:08:19 crc kubenswrapper[4715]: I1204 15:08:19.195227 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3bea4af7-30c3-4278-99d7-01f1c604a89c/glance-log/0.log" Dec 04 15:08:19 crc kubenswrapper[4715]: I1204 15:08:19.491669 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5b798c69c6-g6fvm_8552f780-12e2-4eeb-bbc5-9ef637e9278f/horizon/0.log" Dec 04 15:08:19 crc kubenswrapper[4715]: I1204 15:08:19.590927 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-gn5gv_d394ffb7-2675-4316-8388-cee93c9a37bc/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:19 crc kubenswrapper[4715]: I1204 15:08:19.818583 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6l4js_b45dfcc4-2ef8-4cc0-80e3-14456ad8e0d2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:19 crc kubenswrapper[4715]: I1204 15:08:19.831417 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5b798c69c6-g6fvm_8552f780-12e2-4eeb-bbc5-9ef637e9278f/horizon-log/0.log" Dec 04 15:08:20 crc kubenswrapper[4715]: I1204 15:08:20.030336 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29414341-fhbsf_1e0a90fb-c84b-45ee-ae9f-7975ff939ef0/keystone-cron/0.log" Dec 04 15:08:20 crc kubenswrapper[4715]: I1204 15:08:20.376861 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b0d685d3-1f35-497c-986b-14d7407716d7/kube-state-metrics/0.log" Dec 04 15:08:20 crc kubenswrapper[4715]: I1204 15:08:20.561713 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b79f64b59-wfklg_a412f2fc-2452-41bc-9670-d0db2350f145/keystone-api/0.log" Dec 04 15:08:20 crc kubenswrapper[4715]: I1204 15:08:20.619468 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ndfzt_4a13c152-6be0-4261-ae2f-3b7b3458005a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:21 crc kubenswrapper[4715]: I1204 15:08:21.123943 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58655b5d9c-m5lkw_4e9c813e-1ad1-40fd-b015-a312efff6d9e/neutron-api/0.log" Dec 04 15:08:21 crc kubenswrapper[4715]: I1204 15:08:21.131252 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58655b5d9c-m5lkw_4e9c813e-1ad1-40fd-b015-a312efff6d9e/neutron-httpd/0.log" Dec 04 15:08:21 crc kubenswrapper[4715]: I1204 15:08:21.494704 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xh7vl_e8a7fc4f-4f90-4023-98af-a9613dbce1cb/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.084592 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_02266663-d1a7-4968-ab1e-e4bb6b4a9b21/nova-cell0-conductor-conductor/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.100251 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9ec54636-c541-4292-b5b9-7a6215b8828e/nova-api-log/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.278769 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9ec54636-c541-4292-b5b9-7a6215b8828e/nova-api-api/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.476462 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9701ac27-1f36-4d70-a5ac-2197ea8880f6/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.486921 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0246a68d-8dff-43c6-9efa-fef70ae0919b/nova-cell1-conductor-conductor/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.631192 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-gjztt_df89956f-dd09-4cd1-bc58-d57f555d2870/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:22 crc kubenswrapper[4715]: I1204 15:08:22.830020 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_62b5d096-e3b9-4fae-a697-d5a4b693514d/nova-metadata-log/0.log" Dec 04 15:08:23 crc kubenswrapper[4715]: I1204 15:08:23.505790 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e478a0b5-cea1-4639-a645-9660524ff9bd/mysql-bootstrap/0.log" Dec 04 15:08:23 crc kubenswrapper[4715]: I1204 15:08:23.780602 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e478a0b5-cea1-4639-a645-9660524ff9bd/mysql-bootstrap/0.log" Dec 04 15:08:23 crc kubenswrapper[4715]: I1204 15:08:23.800985 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_23bde1d8-ea21-46b4-8fd6-4fdefc22a6ce/nova-scheduler-scheduler/0.log" Dec 04 15:08:23 crc kubenswrapper[4715]: I1204 15:08:23.881636 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e478a0b5-cea1-4639-a645-9660524ff9bd/galera/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.006237 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7d666280-e175-4f31-8e11-7af2d5e27e5a/mysql-bootstrap/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.219368 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7d666280-e175-4f31-8e11-7af2d5e27e5a/galera/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.254638 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7d666280-e175-4f31-8e11-7af2d5e27e5a/mysql-bootstrap/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.331200 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_62b5d096-e3b9-4fae-a697-d5a4b693514d/nova-metadata-metadata/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.386389 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_454d30ac-7b15-4e7c-8f9c-16c6cfbe5f72/openstackclient/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.466529 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6k4jg_00743321-f084-4e95-b47d-1b21332be3d5/ovn-controller/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.623677 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-tw9p4_44a78956-6eb6-43e2-98f1-2e2c15342fd8/openstack-network-exporter/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.830565 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovsdb-server-init/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.943810 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovsdb-server-init/0.log" Dec 04 15:08:24 crc kubenswrapper[4715]: I1204 15:08:24.979921 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovs-vswitchd/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.034005 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vmdp8_2f0242ef-7fbe-437e-b8c3-6dc6d32c65cf/ovsdb-server/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.237525 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6qf62_d2a5ba42-f0a0-4c65-a87c-fe7ad7d0bcb5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.291543 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_57e4b94e-20df-457c-81e8-9c114b18bbe6/openstack-network-exporter/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.339412 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_57e4b94e-20df-457c-81e8-9c114b18bbe6/ovn-northd/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.474114 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d693b193-c70b-4654-828d-5cd4744bee62/openstack-network-exporter/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.494658 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d693b193-c70b-4654-828d-5cd4744bee62/ovsdbserver-nb/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.696815 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f01e0a29-0003-41df-a5c9-0ede7a249acc/openstack-network-exporter/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.712707 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f01e0a29-0003-41df-a5c9-0ede7a249acc/ovsdbserver-sb/0.log" Dec 04 15:08:25 crc kubenswrapper[4715]: I1204 15:08:25.948361 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65f6dc6d96-gj9d4_aad8392c-fc50-4cd5-af7c-9d186021d336/placement-api/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.047137 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65f6dc6d96-gj9d4_aad8392c-fc50-4cd5-af7c-9d186021d336/placement-log/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.048566 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_45d230ae-3ae5-4d87-8979-bbd1fd8651cd/setup-container/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.229503 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_45d230ae-3ae5-4d87-8979-bbd1fd8651cd/setup-container/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.264198 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_72cfc89d-7970-45e9-be5f-3f5f5d3f520d/setup-container/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.277440 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_45d230ae-3ae5-4d87-8979-bbd1fd8651cd/rabbitmq/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.550236 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_72cfc89d-7970-45e9-be5f-3f5f5d3f520d/setup-container/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.555358 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_72cfc89d-7970-45e9-be5f-3f5f5d3f520d/rabbitmq/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.560894 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dwnhj_baed77e4-c68b-4663-b4c2-0f0de712d963/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.791366 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-plght_9542e2ff-8723-4d1d-b17c-277976018a11/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.836347 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lnjqt_76dfdae9-34be-48e3-a220-4322d19e7260/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:26 crc kubenswrapper[4715]: I1204 15:08:26.987186 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-fpzs7_d300a22e-2124-4f56-a6c7-7f1b4cf5e7e1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.020523 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-4wlm4_5058ed17-7d86-4b90-a41b-668b6407054a/ssh-known-hosts-edpm-deployment/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.373978 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-64bbd97b75-ssj5n_861e33f9-3559-444f-8120-f439948fc774/proxy-server/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.529245 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-64bbd97b75-ssj5n_861e33f9-3559-444f-8120-f439948fc774/proxy-httpd/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.592285 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5r72w_2386cbfc-2eba-4ace-a56a-0e203b7b9229/swift-ring-rebalance/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.769357 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-auditor/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.786185 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-reaper/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.833587 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-replicator/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.884725 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/account-server/0.log" Dec 04 15:08:27 crc kubenswrapper[4715]: I1204 15:08:27.985930 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-replicator/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.126603 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-server/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.132469 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-auditor/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.179098 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/container-updater/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.301278 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-auditor/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.326457 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-expirer/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.368008 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-replicator/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.433203 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-server/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.548748 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/rsync/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.558962 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/object-updater/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.576731 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_548d91b6-2482-4624-843a-ea2898f7beb8/swift-recon-cron/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.760410 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qvgdz_2fa80717-8afa-4d1a-ac4f-4b807445f638/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.838941 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1314b612-6502-4cef-ade7-f05971b781a2/tempest-tests-tempest-tests-runner/0.log" Dec 04 15:08:28 crc kubenswrapper[4715]: I1204 15:08:28.953096 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_212e2726-336d-42d9-942a-36642b0c1c98/test-operator-logs-container/0.log" Dec 04 15:08:29 crc kubenswrapper[4715]: I1204 15:08:29.198812 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cq24p_ee373f57-aa34-4c0a-86d4-5113c10984df/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:08:40 crc kubenswrapper[4715]: I1204 15:08:40.651951 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_91e88f07-d0d0-46d4-9e6a-a973b37b8a72/memcached/0.log" Dec 04 15:08:56 crc kubenswrapper[4715]: I1204 15:08:56.767548 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/util/0.log" Dec 04 15:08:56 crc kubenswrapper[4715]: I1204 15:08:56.919633 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/util/0.log" Dec 04 15:08:56 crc kubenswrapper[4715]: I1204 15:08:56.965320 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/pull/0.log" Dec 04 15:08:56 crc kubenswrapper[4715]: I1204 15:08:56.976427 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/pull/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.140333 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/util/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.163288 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/extract/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.163558 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0f1f8545cb3ca5960856c989219b0b556ab5da78f8fa399ed74d86dcb7x86gw_1d09748d-bd97-4abd-ade6-92e80c61ffbd/pull/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.394563 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-snfrf_d7fc7c62-f1ba-4fcc-9b82-b980988c90cf/kube-rbac-proxy/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.394756 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-snfrf_d7fc7c62-f1ba-4fcc-9b82-b980988c90cf/manager/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.441652 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fdgb9_6efc64e6-15fa-4c60-9a4e-18d326e4c759/kube-rbac-proxy/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.655729 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fdgb9_6efc64e6-15fa-4c60-9a4e-18d326e4c759/manager/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.712024 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-5m69r_4b1a2407-c585-4f22-95db-a2331a69623d/kube-rbac-proxy/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.712829 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-5m69r_4b1a2407-c585-4f22-95db-a2331a69623d/manager/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.931070 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-grcg8_96aa918a-abfb-44f6-bfa3-9849ef0f51e4/kube-rbac-proxy/0.log" Dec 04 15:08:57 crc kubenswrapper[4715]: I1204 15:08:57.978230 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-grcg8_96aa918a-abfb-44f6-bfa3-9849ef0f51e4/manager/0.log" Dec 04 15:08:58 crc kubenswrapper[4715]: I1204 15:08:58.557975 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gv5kg_fbc62f1d-2aa8-4b6b-8693-a16c349461d3/kube-rbac-proxy/0.log" Dec 04 15:08:58 crc kubenswrapper[4715]: I1204 15:08:58.575693 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gv5kg_fbc62f1d-2aa8-4b6b-8693-a16c349461d3/manager/0.log" Dec 04 15:08:58 crc kubenswrapper[4715]: I1204 15:08:58.621073 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5glkm_81510632-c76c-4465-8dec-d8d9c3e8c88d/kube-rbac-proxy/0.log" Dec 04 15:08:58 crc kubenswrapper[4715]: I1204 15:08:58.793763 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5glkm_81510632-c76c-4465-8dec-d8d9c3e8c88d/manager/0.log" Dec 04 15:08:58 crc kubenswrapper[4715]: I1204 15:08:58.798175 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-f8cgr_8a9a0373-b39e-496a-aba7-456bddcf3a65/kube-rbac-proxy/0.log" Dec 04 15:08:58 crc kubenswrapper[4715]: I1204 15:08:58.991767 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-f8cgr_8a9a0373-b39e-496a-aba7-456bddcf3a65/manager/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.028191 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-cw997_ec0c61ca-d902-4703-85bc-87b55736fe78/kube-rbac-proxy/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.086436 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-cw997_ec0c61ca-d902-4703-85bc-87b55736fe78/manager/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.290453 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-rtj2r_9c67b2dc-8ac4-4dd0-af9f-19e55d368092/kube-rbac-proxy/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.550568 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-rtj2r_9c67b2dc-8ac4-4dd0-af9f-19e55d368092/manager/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.637966 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-hhjwx_45a4ad9a-e82a-466a-954b-12127626b577/kube-rbac-proxy/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.691105 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-hhjwx_45a4ad9a-e82a-466a-954b-12127626b577/manager/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.745847 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-xncpr_5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd/kube-rbac-proxy/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.763309 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-xncpr_5b73f1a9-1bf0-4239-84cc-e28bbddcc2bd/manager/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.882530 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6hbp2_3521ae52-a647-4190-8bbd-1100bb31dae5/kube-rbac-proxy/0.log" Dec 04 15:08:59 crc kubenswrapper[4715]: I1204 15:08:59.970647 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6hbp2_3521ae52-a647-4190-8bbd-1100bb31dae5/manager/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.004683 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vnndg_ab0e965e-1d56-4b31-8cf1-6a21892d38ca/kube-rbac-proxy/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.148410 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dg7m6_4238708b-a19f-485f-a8a7-66b58aac168f/kube-rbac-proxy/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.167976 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vnndg_ab0e965e-1d56-4b31-8cf1-6a21892d38ca/manager/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.203054 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dg7m6_4238708b-a19f-485f-a8a7-66b58aac168f/manager/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.296331 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w_73189a74-45fd-4ea8-92ce-bfa038531159/kube-rbac-proxy/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.448742 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4ngf7w_73189a74-45fd-4ea8-92ce-bfa038531159/manager/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.703322 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pnqdc_f7e73b69-ec29-4bdb-9374-25916f8df390/registry-server/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.777594 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6cf7bc7f5d-kzpwj_21b74fe1-833a-4744-b868-6b09c3e93458/operator/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.866713 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-wwc57_84876a00-902b-41f4-af69-f39d588e34fe/kube-rbac-proxy/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.975166 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-wwc57_84876a00-902b-41f4-af69-f39d588e34fe/manager/0.log" Dec 04 15:09:00 crc kubenswrapper[4715]: I1204 15:09:00.979733 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-w2drg_9d367397-9c5c-4ef7-bf95-91d0a453cf04/kube-rbac-proxy/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.157750 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-w2drg_9d367397-9c5c-4ef7-bf95-91d0a453cf04/manager/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.289366 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9g592_43170033-ddda-4fdf-b0e7-8d3b6d15e4f9/operator/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.501810 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qq7m7_6d5d27a6-eba8-4699-a415-e0e8c283d9cf/kube-rbac-proxy/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.556540 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qq7m7_6d5d27a6-eba8-4699-a415-e0e8c283d9cf/manager/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.621858 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-xc2z5_cec0926c-ac5c-4ef2-aefc-a81c7885d810/kube-rbac-proxy/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.628868 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b8c75fd85-txxk5_00ef5230-a1c9-4609-ba71-e90760f29a15/manager/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.763351 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-xc2z5_cec0926c-ac5c-4ef2-aefc-a81c7885d810/manager/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.841682 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-sg77c_88013762-6dd2-4994-b34e-21c89ed5afee/manager/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.872644 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-sg77c_88013762-6dd2-4994-b34e-21c89ed5afee/kube-rbac-proxy/0.log" Dec 04 15:09:01 crc kubenswrapper[4715]: I1204 15:09:01.976983 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-8ztmn_d58523b1-200c-4b54-ab7c-fab0ee9ff571/kube-rbac-proxy/0.log" Dec 04 15:09:02 crc kubenswrapper[4715]: I1204 15:09:02.030308 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-8ztmn_d58523b1-200c-4b54-ab7c-fab0ee9ff571/manager/0.log" Dec 04 15:09:08 crc kubenswrapper[4715]: I1204 15:09:08.761215 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:09:08 crc kubenswrapper[4715]: I1204 15:09:08.761909 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:09:21 crc kubenswrapper[4715]: I1204 15:09:21.856858 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dvrxw_288f2069-b715-4316-ac68-c075546d04c4/control-plane-machine-set-operator/0.log" Dec 04 15:09:22 crc kubenswrapper[4715]: I1204 15:09:22.036131 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c8mr7_0cf11e59-8054-46a5-8a2d-f7c6434a6a91/kube-rbac-proxy/0.log" Dec 04 15:09:22 crc kubenswrapper[4715]: I1204 15:09:22.077374 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c8mr7_0cf11e59-8054-46a5-8a2d-f7c6434a6a91/machine-api-operator/0.log" Dec 04 15:09:35 crc kubenswrapper[4715]: I1204 15:09:35.361622 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-8nphx_70e6d48d-4b22-42ea-9016-a918b6889d78/cert-manager-controller/0.log" Dec 04 15:09:35 crc kubenswrapper[4715]: I1204 15:09:35.508786 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4tnrx_f4374824-e956-48c8-a4e5-52c55c7486e6/cert-manager-cainjector/0.log" Dec 04 15:09:35 crc kubenswrapper[4715]: I1204 15:09:35.597874 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qbkqd_288f92f5-abcf-40e4-937a-93f6f5ba002b/cert-manager-webhook/0.log" Dec 04 15:09:38 crc kubenswrapper[4715]: I1204 15:09:38.758798 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:09:38 crc kubenswrapper[4715]: I1204 15:09:38.759153 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:09:49 crc kubenswrapper[4715]: I1204 15:09:49.098014 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-pg74q_29cbcfcf-ff1d-4605-ac6e-d7c609de5911/nmstate-console-plugin/0.log" Dec 04 15:09:49 crc kubenswrapper[4715]: I1204 15:09:49.245479 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lml75_4165e1e3-7227-4666-a737-14083e64195d/nmstate-handler/0.log" Dec 04 15:09:49 crc kubenswrapper[4715]: I1204 15:09:49.313168 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-nc82h_c4ab73a9-a74b-43dc-bbe5-ad71120feab2/kube-rbac-proxy/0.log" Dec 04 15:09:49 crc kubenswrapper[4715]: I1204 15:09:49.374275 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-nc82h_c4ab73a9-a74b-43dc-bbe5-ad71120feab2/nmstate-metrics/0.log" Dec 04 15:09:49 crc kubenswrapper[4715]: I1204 15:09:49.479444 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-8k4vh_48b345c0-a5f8-4ff5-b696-e147e5ec9753/nmstate-operator/0.log" Dec 04 15:09:49 crc kubenswrapper[4715]: I1204 15:09:49.586958 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-dbfzg_746b0b9b-20ec-42d8-8d9a-4aa11c18a58e/nmstate-webhook/0.log" Dec 04 15:10:03 crc kubenswrapper[4715]: I1204 15:10:03.843205 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-kp8gj_c4a52cd8-e319-4888-8920-a69547290700/kube-rbac-proxy/0.log" Dec 04 15:10:03 crc kubenswrapper[4715]: I1204 15:10:03.919367 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-kp8gj_c4a52cd8-e319-4888-8920-a69547290700/controller/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.081098 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.309229 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.357745 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.358241 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.366725 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.505487 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.555023 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.583251 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.596813 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.782343 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-reloader/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.794676 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-frr-files/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.825788 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/cp-metrics/0.log" Dec 04 15:10:04 crc kubenswrapper[4715]: I1204 15:10:04.853109 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/controller/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.041080 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/frr-metrics/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.045558 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/kube-rbac-proxy-frr/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.125117 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/kube-rbac-proxy/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.292026 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/reloader/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.383426 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-znkxb_d254ca37-56e1-470f-839b-cc07aa846cd3/frr-k8s-webhook-server/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.603244 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d7c69d77f-4r8gn_a57d59fd-4f87-4221-a861-8a4cde8314d6/manager/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.775992 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-78dd5c7fc8-vml6t_0346d3fa-78df-4769-9c41-b1b1ca9682b9/webhook-server/0.log" Dec 04 15:10:05 crc kubenswrapper[4715]: I1204 15:10:05.887813 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mx2g2_1de7de85-bcce-4132-9885-45d7996f6845/kube-rbac-proxy/0.log" Dec 04 15:10:06 crc kubenswrapper[4715]: I1204 15:10:06.429368 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mx2g2_1de7de85-bcce-4132-9885-45d7996f6845/speaker/0.log" Dec 04 15:10:06 crc kubenswrapper[4715]: I1204 15:10:06.449228 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6hp55_a4126a74-57d7-4987-bd7a-091b5e6f3806/frr/0.log" Dec 04 15:10:08 crc kubenswrapper[4715]: I1204 15:10:08.758936 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:10:08 crc kubenswrapper[4715]: I1204 15:10:08.759328 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:10:08 crc kubenswrapper[4715]: I1204 15:10:08.759381 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 15:10:08 crc kubenswrapper[4715]: I1204 15:10:08.760196 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"288ff8ad632dc3f0edd464ac0c1fad0e3b8cc2f3f601075dd1429fe6530ce81f"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:10:08 crc kubenswrapper[4715]: I1204 15:10:08.760319 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://288ff8ad632dc3f0edd464ac0c1fad0e3b8cc2f3f601075dd1429fe6530ce81f" gracePeriod=600 Dec 04 15:10:09 crc kubenswrapper[4715]: I1204 15:10:09.377587 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="288ff8ad632dc3f0edd464ac0c1fad0e3b8cc2f3f601075dd1429fe6530ce81f" exitCode=0 Dec 04 15:10:09 crc kubenswrapper[4715]: I1204 15:10:09.377665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"288ff8ad632dc3f0edd464ac0c1fad0e3b8cc2f3f601075dd1429fe6530ce81f"} Dec 04 15:10:09 crc kubenswrapper[4715]: I1204 15:10:09.378156 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerStarted","Data":"a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d"} Dec 04 15:10:09 crc kubenswrapper[4715]: I1204 15:10:09.378186 4715 scope.go:117] "RemoveContainer" containerID="ad830881ed232704795d9c04cc4833f3895c8189ecacffe173cfb4057a48a375" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.158966 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/util/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.386336 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/util/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.399437 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/pull/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.410265 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/pull/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.596403 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/extract/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.603723 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/util/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.606184 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fc8qnj_b91a857d-7b05-4998-a029-e7576e7de6c9/pull/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.793727 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/util/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.963488 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/util/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.964404 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/pull/0.log" Dec 04 15:10:21 crc kubenswrapper[4715]: I1204 15:10:21.996540 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/pull/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.163715 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/util/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.185406 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/extract/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.203564 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83msp2s_e7c6ddac-1848-4449-8fbe-69d10990b25a/pull/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.344062 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-utilities/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.529666 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-utilities/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.538891 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-content/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.566083 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-content/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.707590 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-utilities/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.714473 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/extract-content/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.870455 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6mjlz_6514d4c8-7452-4255-9e08-276daa522ad0/registry-server/0.log" Dec 04 15:10:22 crc kubenswrapper[4715]: I1204 15:10:22.891205 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-utilities/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.134192 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-content/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.140597 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-utilities/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.211735 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-content/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.470473 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-utilities/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.491537 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/extract-content/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.700548 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-utilities/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.759369 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7wkp2_be7b5959-83d4-44ef-a596-f52b2e2a34f7/marketplace-operator/0.log" Dec 04 15:10:23 crc kubenswrapper[4715]: I1204 15:10:23.956310 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-content/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.016775 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-utilities/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.049102 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-content/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.168574 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h5dgv_26ec7fe9-b52f-4211-a5bf-ab16f43cb7af/registry-server/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.341156 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-content/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.364590 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/extract-utilities/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.529719 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lg6p_0d093b14-52c6-442a-901d-895fa5b9b721/registry-server/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.547547 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-utilities/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.753420 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-content/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.777338 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-utilities/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.827857 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-content/0.log" Dec 04 15:10:24 crc kubenswrapper[4715]: I1204 15:10:24.970105 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-content/0.log" Dec 04 15:10:25 crc kubenswrapper[4715]: I1204 15:10:25.001112 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/extract-utilities/0.log" Dec 04 15:10:25 crc kubenswrapper[4715]: I1204 15:10:25.119856 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b9kpz_5e8b8414-517f-4857-ae4a-c4e131ef8822/registry-server/0.log" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.633885 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8j9v6"] Dec 04 15:11:00 crc kubenswrapper[4715]: E1204 15:11:00.635756 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493fc4e9-b874-4f0a-a14e-c346b482cfa5" containerName="container-00" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.635799 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="493fc4e9-b874-4f0a-a14e-c346b482cfa5" containerName="container-00" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.636583 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="493fc4e9-b874-4f0a-a14e-c346b482cfa5" containerName="container-00" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.639964 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.641857 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8j9v6"] Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.709898 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-utilities\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.709963 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2xdq\" (UniqueName: \"kubernetes.io/projected/3d276848-87b6-43c1-a122-2974297099b9-kube-api-access-d2xdq\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.710092 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-catalog-content\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.811382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2xdq\" (UniqueName: \"kubernetes.io/projected/3d276848-87b6-43c1-a122-2974297099b9-kube-api-access-d2xdq\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.811454 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-catalog-content\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.811759 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-utilities\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.812317 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-utilities\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:00 crc kubenswrapper[4715]: I1204 15:11:00.812680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-catalog-content\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:03 crc kubenswrapper[4715]: I1204 15:11:03.484087 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2xdq\" (UniqueName: \"kubernetes.io/projected/3d276848-87b6-43c1-a122-2974297099b9-kube-api-access-d2xdq\") pod \"redhat-marketplace-8j9v6\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:03 crc kubenswrapper[4715]: I1204 15:11:03.667689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:04 crc kubenswrapper[4715]: I1204 15:11:04.127073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8j9v6"] Dec 04 15:11:04 crc kubenswrapper[4715]: W1204 15:11:04.137552 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d276848_87b6_43c1_a122_2974297099b9.slice/crio-b44db64c1b0f7a5420090f7a4faa9df1a58ce96b3e4bbdd9f28cd52b0fb6ece9 WatchSource:0}: Error finding container b44db64c1b0f7a5420090f7a4faa9df1a58ce96b3e4bbdd9f28cd52b0fb6ece9: Status 404 returned error can't find the container with id b44db64c1b0f7a5420090f7a4faa9df1a58ce96b3e4bbdd9f28cd52b0fb6ece9 Dec 04 15:11:04 crc kubenswrapper[4715]: I1204 15:11:04.905342 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d276848-87b6-43c1-a122-2974297099b9" containerID="64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7" exitCode=0 Dec 04 15:11:04 crc kubenswrapper[4715]: I1204 15:11:04.905395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8j9v6" event={"ID":"3d276848-87b6-43c1-a122-2974297099b9","Type":"ContainerDied","Data":"64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7"} Dec 04 15:11:04 crc kubenswrapper[4715]: I1204 15:11:04.905668 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8j9v6" event={"ID":"3d276848-87b6-43c1-a122-2974297099b9","Type":"ContainerStarted","Data":"b44db64c1b0f7a5420090f7a4faa9df1a58ce96b3e4bbdd9f28cd52b0fb6ece9"} Dec 04 15:11:06 crc kubenswrapper[4715]: I1204 15:11:06.942969 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d276848-87b6-43c1-a122-2974297099b9" containerID="06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f" exitCode=0 Dec 04 15:11:06 crc kubenswrapper[4715]: I1204 15:11:06.943382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8j9v6" event={"ID":"3d276848-87b6-43c1-a122-2974297099b9","Type":"ContainerDied","Data":"06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f"} Dec 04 15:11:07 crc kubenswrapper[4715]: I1204 15:11:07.955098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8j9v6" event={"ID":"3d276848-87b6-43c1-a122-2974297099b9","Type":"ContainerStarted","Data":"63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b"} Dec 04 15:11:07 crc kubenswrapper[4715]: I1204 15:11:07.987904 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8j9v6" podStartSLOduration=5.573663202 podStartE2EDuration="7.987880548s" podCreationTimestamp="2025-12-04 15:11:00 +0000 UTC" firstStartedPulling="2025-12-04 15:11:04.907360156 +0000 UTC m=+4421.976078371" lastFinishedPulling="2025-12-04 15:11:07.321577502 +0000 UTC m=+4424.390295717" observedRunningTime="2025-12-04 15:11:07.975587448 +0000 UTC m=+4425.044305673" watchObservedRunningTime="2025-12-04 15:11:07.987880548 +0000 UTC m=+4425.056598763" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.373544 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-grghd"] Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.378057 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-grghd"] Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.378199 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.529658 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-utilities\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.529714 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-catalog-content\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.529781 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st7pn\" (UniqueName: \"kubernetes.io/projected/08be25b0-703f-4634-ad9c-959a04e39368-kube-api-access-st7pn\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.632323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-utilities\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.632379 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-catalog-content\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.632431 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st7pn\" (UniqueName: \"kubernetes.io/projected/08be25b0-703f-4634-ad9c-959a04e39368-kube-api-access-st7pn\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.632934 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-utilities\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.633167 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-catalog-content\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.665647 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st7pn\" (UniqueName: \"kubernetes.io/projected/08be25b0-703f-4634-ad9c-959a04e39368-kube-api-access-st7pn\") pod \"certified-operators-grghd\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:10 crc kubenswrapper[4715]: I1204 15:11:10.741221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:11 crc kubenswrapper[4715]: I1204 15:11:11.254927 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-grghd"] Dec 04 15:11:12 crc kubenswrapper[4715]: I1204 15:11:12.000243 4715 generic.go:334] "Generic (PLEG): container finished" podID="08be25b0-703f-4634-ad9c-959a04e39368" containerID="bb615319ac483591f641070e74a5b5375dd793daa300af53727189ea6b4aab37" exitCode=0 Dec 04 15:11:12 crc kubenswrapper[4715]: I1204 15:11:12.000297 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerDied","Data":"bb615319ac483591f641070e74a5b5375dd793daa300af53727189ea6b4aab37"} Dec 04 15:11:12 crc kubenswrapper[4715]: I1204 15:11:12.000552 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerStarted","Data":"2120b3dfd0030b051169a0b75f8a2cb8838c0992c1a97f619190851237d70b62"} Dec 04 15:11:13 crc kubenswrapper[4715]: I1204 15:11:13.668621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:13 crc kubenswrapper[4715]: I1204 15:11:13.670240 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:13 crc kubenswrapper[4715]: I1204 15:11:13.721790 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:14 crc kubenswrapper[4715]: I1204 15:11:14.017295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerStarted","Data":"6334f40773c54a5b2bcb37f7d2323410f0f425e6864941c3603a38b50934e9a3"} Dec 04 15:11:14 crc kubenswrapper[4715]: I1204 15:11:14.069489 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:15 crc kubenswrapper[4715]: I1204 15:11:15.028167 4715 generic.go:334] "Generic (PLEG): container finished" podID="08be25b0-703f-4634-ad9c-959a04e39368" containerID="6334f40773c54a5b2bcb37f7d2323410f0f425e6864941c3603a38b50934e9a3" exitCode=0 Dec 04 15:11:15 crc kubenswrapper[4715]: I1204 15:11:15.029834 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerDied","Data":"6334f40773c54a5b2bcb37f7d2323410f0f425e6864941c3603a38b50934e9a3"} Dec 04 15:11:15 crc kubenswrapper[4715]: I1204 15:11:15.346183 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8j9v6"] Dec 04 15:11:16 crc kubenswrapper[4715]: I1204 15:11:16.043457 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerStarted","Data":"6b961efab8551577a3eb409cc7c2de972b6fc37b2e3cf89b1bbdb4b469859c75"} Dec 04 15:11:16 crc kubenswrapper[4715]: I1204 15:11:16.072250 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-grghd" podStartSLOduration=2.672214327 podStartE2EDuration="6.072222402s" podCreationTimestamp="2025-12-04 15:11:10 +0000 UTC" firstStartedPulling="2025-12-04 15:11:12.002405849 +0000 UTC m=+4429.071124064" lastFinishedPulling="2025-12-04 15:11:15.402413914 +0000 UTC m=+4432.471132139" observedRunningTime="2025-12-04 15:11:16.061836662 +0000 UTC m=+4433.130554887" watchObservedRunningTime="2025-12-04 15:11:16.072222402 +0000 UTC m=+4433.140940627" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.054695 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8j9v6" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="registry-server" containerID="cri-o://63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b" gracePeriod=2 Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.527686 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.726165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-utilities\") pod \"3d276848-87b6-43c1-a122-2974297099b9\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.726290 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2xdq\" (UniqueName: \"kubernetes.io/projected/3d276848-87b6-43c1-a122-2974297099b9-kube-api-access-d2xdq\") pod \"3d276848-87b6-43c1-a122-2974297099b9\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.726432 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-catalog-content\") pod \"3d276848-87b6-43c1-a122-2974297099b9\" (UID: \"3d276848-87b6-43c1-a122-2974297099b9\") " Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.727367 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-utilities" (OuterVolumeSpecName: "utilities") pod "3d276848-87b6-43c1-a122-2974297099b9" (UID: "3d276848-87b6-43c1-a122-2974297099b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.736456 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d276848-87b6-43c1-a122-2974297099b9-kube-api-access-d2xdq" (OuterVolumeSpecName: "kube-api-access-d2xdq") pod "3d276848-87b6-43c1-a122-2974297099b9" (UID: "3d276848-87b6-43c1-a122-2974297099b9"). InnerVolumeSpecName "kube-api-access-d2xdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.749564 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d276848-87b6-43c1-a122-2974297099b9" (UID: "3d276848-87b6-43c1-a122-2974297099b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.828605 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2xdq\" (UniqueName: \"kubernetes.io/projected/3d276848-87b6-43c1-a122-2974297099b9-kube-api-access-d2xdq\") on node \"crc\" DevicePath \"\"" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.828669 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:11:17 crc kubenswrapper[4715]: I1204 15:11:17.828684 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d276848-87b6-43c1-a122-2974297099b9-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.065288 4715 generic.go:334] "Generic (PLEG): container finished" podID="3d276848-87b6-43c1-a122-2974297099b9" containerID="63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b" exitCode=0 Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.065352 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8j9v6" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.065370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8j9v6" event={"ID":"3d276848-87b6-43c1-a122-2974297099b9","Type":"ContainerDied","Data":"63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b"} Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.076089 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8j9v6" event={"ID":"3d276848-87b6-43c1-a122-2974297099b9","Type":"ContainerDied","Data":"b44db64c1b0f7a5420090f7a4faa9df1a58ce96b3e4bbdd9f28cd52b0fb6ece9"} Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.076139 4715 scope.go:117] "RemoveContainer" containerID="63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.104971 4715 scope.go:117] "RemoveContainer" containerID="06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.113866 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8j9v6"] Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.121502 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8j9v6"] Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.128295 4715 scope.go:117] "RemoveContainer" containerID="64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.173906 4715 scope.go:117] "RemoveContainer" containerID="63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b" Dec 04 15:11:18 crc kubenswrapper[4715]: E1204 15:11:18.174540 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b\": container with ID starting with 63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b not found: ID does not exist" containerID="63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.174594 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b"} err="failed to get container status \"63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b\": rpc error: code = NotFound desc = could not find container \"63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b\": container with ID starting with 63f0dd2811c343a4ed3c393986e6444da070345567285a697febecb47dd9769b not found: ID does not exist" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.174631 4715 scope.go:117] "RemoveContainer" containerID="06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f" Dec 04 15:11:18 crc kubenswrapper[4715]: E1204 15:11:18.175093 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f\": container with ID starting with 06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f not found: ID does not exist" containerID="06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.175258 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f"} err="failed to get container status \"06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f\": rpc error: code = NotFound desc = could not find container \"06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f\": container with ID starting with 06932aa694d8cab6b578b0f9a3063e3e564e9c78f2265d556fd42859b0fc9b0f not found: ID does not exist" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.175279 4715 scope.go:117] "RemoveContainer" containerID="64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7" Dec 04 15:11:18 crc kubenswrapper[4715]: E1204 15:11:18.175670 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7\": container with ID starting with 64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7 not found: ID does not exist" containerID="64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7" Dec 04 15:11:18 crc kubenswrapper[4715]: I1204 15:11:18.175702 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7"} err="failed to get container status \"64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7\": rpc error: code = NotFound desc = could not find container \"64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7\": container with ID starting with 64d88f0634f78976d6a648b6aab5c0be4615f73520903bbc63605fd97486d4f7 not found: ID does not exist" Dec 04 15:11:19 crc kubenswrapper[4715]: I1204 15:11:19.193789 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d276848-87b6-43c1-a122-2974297099b9" path="/var/lib/kubelet/pods/3d276848-87b6-43c1-a122-2974297099b9/volumes" Dec 04 15:11:20 crc kubenswrapper[4715]: I1204 15:11:20.742627 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:20 crc kubenswrapper[4715]: I1204 15:11:20.742994 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:20 crc kubenswrapper[4715]: I1204 15:11:20.787182 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:21 crc kubenswrapper[4715]: I1204 15:11:21.194248 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:22 crc kubenswrapper[4715]: I1204 15:11:22.127646 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-grghd"] Dec 04 15:11:23 crc kubenswrapper[4715]: I1204 15:11:23.162158 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-grghd" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="registry-server" containerID="cri-o://6b961efab8551577a3eb409cc7c2de972b6fc37b2e3cf89b1bbdb4b469859c75" gracePeriod=2 Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.192616 4715 generic.go:334] "Generic (PLEG): container finished" podID="08be25b0-703f-4634-ad9c-959a04e39368" containerID="6b961efab8551577a3eb409cc7c2de972b6fc37b2e3cf89b1bbdb4b469859c75" exitCode=0 Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.192761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerDied","Data":"6b961efab8551577a3eb409cc7c2de972b6fc37b2e3cf89b1bbdb4b469859c75"} Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.429469 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.544064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-catalog-content\") pod \"08be25b0-703f-4634-ad9c-959a04e39368\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.544162 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-utilities\") pod \"08be25b0-703f-4634-ad9c-959a04e39368\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.544190 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st7pn\" (UniqueName: \"kubernetes.io/projected/08be25b0-703f-4634-ad9c-959a04e39368-kube-api-access-st7pn\") pod \"08be25b0-703f-4634-ad9c-959a04e39368\" (UID: \"08be25b0-703f-4634-ad9c-959a04e39368\") " Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.545124 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-utilities" (OuterVolumeSpecName: "utilities") pod "08be25b0-703f-4634-ad9c-959a04e39368" (UID: "08be25b0-703f-4634-ad9c-959a04e39368"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.553829 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08be25b0-703f-4634-ad9c-959a04e39368-kube-api-access-st7pn" (OuterVolumeSpecName: "kube-api-access-st7pn") pod "08be25b0-703f-4634-ad9c-959a04e39368" (UID: "08be25b0-703f-4634-ad9c-959a04e39368"). InnerVolumeSpecName "kube-api-access-st7pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.646796 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:11:24 crc kubenswrapper[4715]: I1204 15:11:24.646848 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st7pn\" (UniqueName: \"kubernetes.io/projected/08be25b0-703f-4634-ad9c-959a04e39368-kube-api-access-st7pn\") on node \"crc\" DevicePath \"\"" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.205562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grghd" event={"ID":"08be25b0-703f-4634-ad9c-959a04e39368","Type":"ContainerDied","Data":"2120b3dfd0030b051169a0b75f8a2cb8838c0992c1a97f619190851237d70b62"} Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.205897 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grghd" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.206704 4715 scope.go:117] "RemoveContainer" containerID="6b961efab8551577a3eb409cc7c2de972b6fc37b2e3cf89b1bbdb4b469859c75" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.229150 4715 scope.go:117] "RemoveContainer" containerID="6334f40773c54a5b2bcb37f7d2323410f0f425e6864941c3603a38b50934e9a3" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.310981 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08be25b0-703f-4634-ad9c-959a04e39368" (UID: "08be25b0-703f-4634-ad9c-959a04e39368"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.360846 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be25b0-703f-4634-ad9c-959a04e39368-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.503476 4715 scope.go:117] "RemoveContainer" containerID="bb615319ac483591f641070e74a5b5375dd793daa300af53727189ea6b4aab37" Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.642594 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-grghd"] Dec 04 15:11:25 crc kubenswrapper[4715]: I1204 15:11:25.652827 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-grghd"] Dec 04 15:11:27 crc kubenswrapper[4715]: I1204 15:11:27.192316 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08be25b0-703f-4634-ad9c-959a04e39368" path="/var/lib/kubelet/pods/08be25b0-703f-4634-ad9c-959a04e39368/volumes" Dec 04 15:12:14 crc kubenswrapper[4715]: I1204 15:12:14.684133 4715 generic.go:334] "Generic (PLEG): container finished" podID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerID="005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d" exitCode=0 Dec 04 15:12:14 crc kubenswrapper[4715]: I1204 15:12:14.684234 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" event={"ID":"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8","Type":"ContainerDied","Data":"005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d"} Dec 04 15:12:14 crc kubenswrapper[4715]: I1204 15:12:14.685466 4715 scope.go:117] "RemoveContainer" containerID="005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d" Dec 04 15:12:15 crc kubenswrapper[4715]: I1204 15:12:15.591880 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7z8qr_must-gather-qgp2k_fc242cfd-fb20-49c1-8812-bc13d5ccb2f8/gather/0.log" Dec 04 15:12:25 crc kubenswrapper[4715]: I1204 15:12:25.978836 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7z8qr/must-gather-qgp2k"] Dec 04 15:12:25 crc kubenswrapper[4715]: I1204 15:12:25.979713 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="copy" containerID="cri-o://e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d" gracePeriod=2 Dec 04 15:12:25 crc kubenswrapper[4715]: I1204 15:12:25.994893 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7z8qr/must-gather-qgp2k"] Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.448293 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7z8qr_must-gather-qgp2k_fc242cfd-fb20-49c1-8812-bc13d5ccb2f8/copy/0.log" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.449078 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.551981 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-must-gather-output\") pod \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.552277 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxwjl\" (UniqueName: \"kubernetes.io/projected/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-kube-api-access-hxwjl\") pod \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\" (UID: \"fc242cfd-fb20-49c1-8812-bc13d5ccb2f8\") " Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.705146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" (UID: "fc242cfd-fb20-49c1-8812-bc13d5ccb2f8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.756496 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.799427 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7z8qr_must-gather-qgp2k_fc242cfd-fb20-49c1-8812-bc13d5ccb2f8/copy/0.log" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.799932 4715 generic.go:334] "Generic (PLEG): container finished" podID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerID="e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d" exitCode=143 Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.799967 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7z8qr/must-gather-qgp2k" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.799992 4715 scope.go:117] "RemoveContainer" containerID="e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d" Dec 04 15:12:26 crc kubenswrapper[4715]: I1204 15:12:26.827812 4715 scope.go:117] "RemoveContainer" containerID="005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.072901 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-kube-api-access-hxwjl" (OuterVolumeSpecName: "kube-api-access-hxwjl") pod "fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" (UID: "fc242cfd-fb20-49c1-8812-bc13d5ccb2f8"). InnerVolumeSpecName "kube-api-access-hxwjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.073830 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxwjl\" (UniqueName: \"kubernetes.io/projected/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8-kube-api-access-hxwjl\") on node \"crc\" DevicePath \"\"" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.139517 4715 scope.go:117] "RemoveContainer" containerID="e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d" Dec 04 15:12:27 crc kubenswrapper[4715]: E1204 15:12:27.140288 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d\": container with ID starting with e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d not found: ID does not exist" containerID="e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.140337 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d"} err="failed to get container status \"e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d\": rpc error: code = NotFound desc = could not find container \"e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d\": container with ID starting with e0bb56725fe856f84a004a00feb5415b8efc289515ed1e9350267ac243dbd54d not found: ID does not exist" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.140369 4715 scope.go:117] "RemoveContainer" containerID="005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d" Dec 04 15:12:27 crc kubenswrapper[4715]: E1204 15:12:27.140745 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d\": container with ID starting with 005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d not found: ID does not exist" containerID="005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.140775 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d"} err="failed to get container status \"005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d\": rpc error: code = NotFound desc = could not find container \"005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d\": container with ID starting with 005b27620efb1b996e026ec7464b000576daa00bb3330309aedcb27f024b694d not found: ID does not exist" Dec 04 15:12:27 crc kubenswrapper[4715]: I1204 15:12:27.192725 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" path="/var/lib/kubelet/pods/fc242cfd-fb20-49c1-8812-bc13d5ccb2f8/volumes" Dec 04 15:12:38 crc kubenswrapper[4715]: I1204 15:12:38.759083 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:12:38 crc kubenswrapper[4715]: I1204 15:12:38.759611 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:13:08 crc kubenswrapper[4715]: I1204 15:13:08.758894 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:13:08 crc kubenswrapper[4715]: I1204 15:13:08.759538 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:13:38 crc kubenswrapper[4715]: I1204 15:13:38.758176 4715 patch_prober.go:28] interesting pod/machine-config-daemon-tblzf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:13:38 crc kubenswrapper[4715]: I1204 15:13:38.758720 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:13:38 crc kubenswrapper[4715]: I1204 15:13:38.758773 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" Dec 04 15:13:38 crc kubenswrapper[4715]: I1204 15:13:38.759612 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d"} pod="openshift-machine-config-operator/machine-config-daemon-tblzf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:13:38 crc kubenswrapper[4715]: I1204 15:13:38.759749 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerName="machine-config-daemon" containerID="cri-o://a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" gracePeriod=600 Dec 04 15:13:38 crc kubenswrapper[4715]: E1204 15:13:38.885792 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:13:39 crc kubenswrapper[4715]: I1204 15:13:39.147007 4715 generic.go:334] "Generic (PLEG): container finished" podID="4d5ee7c9-f82d-4848-b21f-e529d843456e" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" exitCode=0 Dec 04 15:13:39 crc kubenswrapper[4715]: I1204 15:13:39.147067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" event={"ID":"4d5ee7c9-f82d-4848-b21f-e529d843456e","Type":"ContainerDied","Data":"a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d"} Dec 04 15:13:39 crc kubenswrapper[4715]: I1204 15:13:39.147127 4715 scope.go:117] "RemoveContainer" containerID="288ff8ad632dc3f0edd464ac0c1fad0e3b8cc2f3f601075dd1429fe6530ce81f" Dec 04 15:13:39 crc kubenswrapper[4715]: I1204 15:13:39.147967 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:13:39 crc kubenswrapper[4715]: E1204 15:13:39.148326 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:13:53 crc kubenswrapper[4715]: I1204 15:13:53.187019 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:13:53 crc kubenswrapper[4715]: E1204 15:13:53.187762 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:14:08 crc kubenswrapper[4715]: I1204 15:14:08.180675 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:14:08 crc kubenswrapper[4715]: E1204 15:14:08.181292 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:14:19 crc kubenswrapper[4715]: I1204 15:14:19.180904 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:14:19 crc kubenswrapper[4715]: E1204 15:14:19.181742 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:14:30 crc kubenswrapper[4715]: I1204 15:14:30.181257 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:14:30 crc kubenswrapper[4715]: E1204 15:14:30.182665 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:14:41 crc kubenswrapper[4715]: I1204 15:14:41.181636 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:14:41 crc kubenswrapper[4715]: E1204 15:14:41.182406 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:14:46 crc kubenswrapper[4715]: I1204 15:14:46.284233 4715 scope.go:117] "RemoveContainer" containerID="5335f11eaf05cc58a4f7af51f080de60183870a634d96d8f161210d64cbeb785" Dec 04 15:14:52 crc kubenswrapper[4715]: I1204 15:14:52.181955 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:14:52 crc kubenswrapper[4715]: E1204 15:14:52.183009 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.177895 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr"] Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.180062 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="copy" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.180189 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="copy" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.180274 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.180345 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.180466 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.181431 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.181724 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="extract-utilities" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.181813 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="extract-utilities" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.181887 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="extract-content" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.181994 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="extract-content" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.182106 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="extract-utilities" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.182185 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="extract-utilities" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.182277 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="gather" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.182354 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="gather" Dec 04 15:15:00 crc kubenswrapper[4715]: E1204 15:15:00.182442 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="extract-content" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.182518 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="extract-content" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.183264 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="copy" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.183397 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="08be25b0-703f-4634-ad9c-959a04e39368" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.183605 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc242cfd-fb20-49c1-8812-bc13d5ccb2f8" containerName="gather" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.183686 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d276848-87b6-43c1-a122-2974297099b9" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.184781 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.188625 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.188651 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.192624 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr"] Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.352883 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/925dd98f-9ad3-422e-ab82-f5df913537ea-config-volume\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.353411 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/925dd98f-9ad3-422e-ab82-f5df913537ea-secret-volume\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.354235 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq7sx\" (UniqueName: \"kubernetes.io/projected/925dd98f-9ad3-422e-ab82-f5df913537ea-kube-api-access-pq7sx\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.455671 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq7sx\" (UniqueName: \"kubernetes.io/projected/925dd98f-9ad3-422e-ab82-f5df913537ea-kube-api-access-pq7sx\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.455743 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/925dd98f-9ad3-422e-ab82-f5df913537ea-config-volume\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.455829 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/925dd98f-9ad3-422e-ab82-f5df913537ea-secret-volume\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.457860 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/925dd98f-9ad3-422e-ab82-f5df913537ea-config-volume\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.462600 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/925dd98f-9ad3-422e-ab82-f5df913537ea-secret-volume\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.475822 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq7sx\" (UniqueName: \"kubernetes.io/projected/925dd98f-9ad3-422e-ab82-f5df913537ea-kube-api-access-pq7sx\") pod \"collect-profiles-29414355-g9hrr\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.526133 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:00 crc kubenswrapper[4715]: I1204 15:15:00.978918 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr"] Dec 04 15:15:01 crc kubenswrapper[4715]: I1204 15:15:01.912203 4715 generic.go:334] "Generic (PLEG): container finished" podID="925dd98f-9ad3-422e-ab82-f5df913537ea" containerID="35fc2e9a1496014c9d38fa75e3d454b6871db79afb3696119204c74f23491d6b" exitCode=0 Dec 04 15:15:01 crc kubenswrapper[4715]: I1204 15:15:01.912352 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" event={"ID":"925dd98f-9ad3-422e-ab82-f5df913537ea","Type":"ContainerDied","Data":"35fc2e9a1496014c9d38fa75e3d454b6871db79afb3696119204c74f23491d6b"} Dec 04 15:15:01 crc kubenswrapper[4715]: I1204 15:15:01.912524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" event={"ID":"925dd98f-9ad3-422e-ab82-f5df913537ea","Type":"ContainerStarted","Data":"4a12da13ea1687818c0c435c43be6ac596bb351c26312c777908d79379992288"} Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.249214 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.413309 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq7sx\" (UniqueName: \"kubernetes.io/projected/925dd98f-9ad3-422e-ab82-f5df913537ea-kube-api-access-pq7sx\") pod \"925dd98f-9ad3-422e-ab82-f5df913537ea\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.413422 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/925dd98f-9ad3-422e-ab82-f5df913537ea-config-volume\") pod \"925dd98f-9ad3-422e-ab82-f5df913537ea\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.413471 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/925dd98f-9ad3-422e-ab82-f5df913537ea-secret-volume\") pod \"925dd98f-9ad3-422e-ab82-f5df913537ea\" (UID: \"925dd98f-9ad3-422e-ab82-f5df913537ea\") " Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.414305 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925dd98f-9ad3-422e-ab82-f5df913537ea-config-volume" (OuterVolumeSpecName: "config-volume") pod "925dd98f-9ad3-422e-ab82-f5df913537ea" (UID: "925dd98f-9ad3-422e-ab82-f5df913537ea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.419241 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925dd98f-9ad3-422e-ab82-f5df913537ea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "925dd98f-9ad3-422e-ab82-f5df913537ea" (UID: "925dd98f-9ad3-422e-ab82-f5df913537ea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.419365 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925dd98f-9ad3-422e-ab82-f5df913537ea-kube-api-access-pq7sx" (OuterVolumeSpecName: "kube-api-access-pq7sx") pod "925dd98f-9ad3-422e-ab82-f5df913537ea" (UID: "925dd98f-9ad3-422e-ab82-f5df913537ea"). InnerVolumeSpecName "kube-api-access-pq7sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.515295 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq7sx\" (UniqueName: \"kubernetes.io/projected/925dd98f-9ad3-422e-ab82-f5df913537ea-kube-api-access-pq7sx\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.515336 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/925dd98f-9ad3-422e-ab82-f5df913537ea-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.515345 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/925dd98f-9ad3-422e-ab82-f5df913537ea-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.930010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" event={"ID":"925dd98f-9ad3-422e-ab82-f5df913537ea","Type":"ContainerDied","Data":"4a12da13ea1687818c0c435c43be6ac596bb351c26312c777908d79379992288"} Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.930078 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a12da13ea1687818c0c435c43be6ac596bb351c26312c777908d79379992288" Dec 04 15:15:03 crc kubenswrapper[4715]: I1204 15:15:03.930087 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-g9hrr" Dec 04 15:15:04 crc kubenswrapper[4715]: I1204 15:15:04.319684 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h"] Dec 04 15:15:04 crc kubenswrapper[4715]: I1204 15:15:04.328108 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-mch6h"] Dec 04 15:15:05 crc kubenswrapper[4715]: I1204 15:15:05.181085 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:15:05 crc kubenswrapper[4715]: E1204 15:15:05.181433 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:15:05 crc kubenswrapper[4715]: I1204 15:15:05.193327 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d505f96-7810-4021-9fd2-5abd46d6ddc6" path="/var/lib/kubelet/pods/8d505f96-7810-4021-9fd2-5abd46d6ddc6/volumes" Dec 04 15:15:20 crc kubenswrapper[4715]: I1204 15:15:20.181346 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:15:20 crc kubenswrapper[4715]: E1204 15:15:20.181975 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:15:32 crc kubenswrapper[4715]: I1204 15:15:32.181211 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:15:32 crc kubenswrapper[4715]: E1204 15:15:32.183032 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:15:44 crc kubenswrapper[4715]: I1204 15:15:44.180263 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:15:44 crc kubenswrapper[4715]: E1204 15:15:44.180944 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:15:46 crc kubenswrapper[4715]: I1204 15:15:46.331886 4715 scope.go:117] "RemoveContainer" containerID="0b98bc12264fb0cb65b8a4b9ada1c324327d832201e8e8ae163580aacb253eec" Dec 04 15:15:55 crc kubenswrapper[4715]: I1204 15:15:55.183142 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:15:55 crc kubenswrapper[4715]: E1204 15:15:55.185529 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" Dec 04 15:16:10 crc kubenswrapper[4715]: I1204 15:16:10.180593 4715 scope.go:117] "RemoveContainer" containerID="a226c5535a52017658e6737b4ab5ca7ded1fa6e67ca7c3f74fec68662906985d" Dec 04 15:16:10 crc kubenswrapper[4715]: E1204 15:16:10.181388 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblzf_openshift-machine-config-operator(4d5ee7c9-f82d-4848-b21f-e529d843456e)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblzf" podUID="4d5ee7c9-f82d-4848-b21f-e529d843456e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114322706024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114322707017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114311202016473 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114311203015444 5ustar corecore